WO2021243828A1 - Procédé et appareil de traitement de texte fondés sur un apprentissage automatique et dispositif informatique et support - Google Patents

Procédé et appareil de traitement de texte fondés sur un apprentissage automatique et dispositif informatique et support Download PDF

Info

Publication number
WO2021243828A1
WO2021243828A1 PCT/CN2020/103784 CN2020103784W WO2021243828A1 WO 2021243828 A1 WO2021243828 A1 WO 2021243828A1 CN 2020103784 W CN2020103784 W CN 2020103784W WO 2021243828 A1 WO2021243828 A1 WO 2021243828A1
Authority
WO
WIPO (PCT)
Prior art keywords
information
standard
answer
data
answer data
Prior art date
Application number
PCT/CN2020/103784
Other languages
English (en)
Chinese (zh)
Inventor
柳阳
喻宁
郑喜民
梁关林
Original Assignee
平安国际智慧城市科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安国际智慧城市科技股份有限公司 filed Critical 平安国际智慧城市科技股份有限公司
Publication of WO2021243828A1 publication Critical patent/WO2021243828A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/33Querying
    • G06F16/332Query formulation
    • G06F16/3329Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/30Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
    • G06F16/35Clustering; Classification
    • G06F16/353Clustering; Classification into predefined classes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • G06F18/24155Bayesian classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods

Definitions

  • This application relates to the field of intelligent decision-making in artificial intelligence, and in particular to a text processing method, device, computer equipment and medium based on machine learning.
  • Machine Reading Comprehension has become a new hot spot in the field of artificial intelligence research and application. Its main function is to read and understand a given article or Context, automatically give answers to related questions.
  • the traditional method of machine reading comprehension mainly adopts the method of determining the correct answer based on similarity or correlation. This type of method determines the correct answer by calculating the most similarity or correlation between the sentence of the option and the background material.
  • Sentences that are semantically equivalent are often expressed in different forms of syntactic structure.
  • the embodiments of the present application provide a text processing method, device, computer equipment, and medium based on machine learning to solve the problem of low accuracy of answers obtained by machine reading.
  • a text processing method based on machine learning including:
  • the standard material information, the standard question information, and the corresponding question type are input into a preset target machine reading comprehension model for prediction, and initial answer information is obtained.
  • the initial answer information includes multiple evaluation data information and Problem-solving idea information corresponding to the standard topic information, wherein the target machine reading comprehension model is obtained by training using a convolutional neural network-pre-training language model;
  • the final evaluation data is determined from a plurality of the evaluation data information according to the problem-solving idea information, and the final evaluation data and the problem-solving idea information are recorded as target answer information in a preset integration manner.
  • a text processing device based on machine learning includes:
  • the preprocessing module is used to obtain the answer data to be processed, and preprocess the answer data to be processed to obtain standard answer data, where the standard answer data includes standard material information and standard question information;
  • the first input module is configured to input the standard question information in the standard answer data into a preset answer classification model to obtain the question type of the standard question information;
  • the prediction module is used to input the standard material information, the standard question information, and the corresponding question type into a preset target machine reading comprehension model for prediction to obtain initial answer information, where the initial answer information includes multiple Pieces of evaluation data information and problem-solving ideas information corresponding to the standard topic information, wherein the target machine reading comprehension model is obtained by training using a convolutional neural network-pre-training language model;
  • the determining module is configured to determine final evaluation data from a plurality of the evaluation data information according to the problem-solving idea information, and record the final evaluation data and the problem-solving idea information as a target answer in a preset integration manner information.
  • a computer device includes a memory and a processor, the processor and the memory are connected to each other, wherein the memory is used to store a computer program, the computer program includes program instructions, and the processor is used to execute the The program instructions of the memory, wherein:
  • the standard material information, the standard question information, and the corresponding question type are input into a preset target machine reading comprehension model for prediction, and initial answer information is obtained.
  • the initial answer information includes multiple evaluation data information and Problem-solving idea information corresponding to the standard topic information, wherein the target machine reading comprehension model is obtained by training using a convolutional neural network-pre-training language model;
  • the final evaluation data is determined from a plurality of the evaluation data information according to the problem-solving idea information, and the final evaluation data and the problem-solving idea information are recorded as target answer information in a preset integration manner.
  • a computer-readable storage medium stores a computer program
  • the computer program includes program instructions, and when the program instructions are executed by a processor, they are used to implement the following steps:
  • the standard material information, the standard question information, and the corresponding question type are input into a preset target machine reading comprehension model for prediction, and initial answer information is obtained.
  • the initial answer information includes multiple evaluation data information and Problem-solving idea information corresponding to the standard topic information, wherein the target machine reading comprehension model is obtained by training using a convolutional neural network-pre-training language model;
  • the final evaluation data is determined from a plurality of the evaluation data information according to the problem-solving idea information, and the final evaluation data and the problem-solving idea information are recorded as target answer information in a preset integration manner.
  • the above-mentioned text processing methods, devices, computer equipment and media based on machine learning obtain standard answer data by taking the answer data to be processed and preprocessing the answer data to be processed.
  • the standard answer data includes standard material information and standard question information;
  • the standard question information in the answer data is input into the preset answer classification model to obtain the question type of the standard question information;
  • the standard material information, standard question information and the corresponding question type are input into the preset target machine reading comprehension model Predict and obtain initial answer information.
  • the initial answer information includes multiple evaluation data information and problem-solving ideas information corresponding to the standard problem information.
  • the target machine reading comprehension model is obtained by training with a convolutional neural network-pre-training language model ⁇ ; According to the problem-solving idea information, the final evaluation data is determined from a plurality of the evaluation data information, and the final evaluation data and the problem-solving idea information are recorded as the target answer information in a preset integration mode; by using a convolutional neural network- The target machine reading comprehension model trained by the pre-training language model predicts the answer to the answer data to be processed, and obtains the target answer information that contains both the evaluation data information and the corresponding problem-solving idea information; thereby further improving the accuracy of the answer obtained by machine reading And the real meaning plays a role in assisting teaching/learning.
  • FIG. 1 is a schematic diagram of an application environment of a text processing method based on machine learning in an embodiment of the present application
  • FIG. 2 is a flowchart of a text processing method based on machine learning in an embodiment of the present application
  • FIG. 3 is another flowchart of a text processing method based on machine learning in an embodiment of the present application
  • FIG. 5 is another flowchart of a text processing method based on machine learning in an embodiment of the present application.
  • Fig. 6 is another flowchart of a text processing method based on machine learning in an embodiment of the present application.
  • FIG. 7 is a functional block diagram of a text processing device based on machine learning in an embodiment of the present application.
  • FIG. 8 is another functional block diagram of a text processing device based on machine learning in an embodiment of the present application.
  • FIG. 9 is another functional block diagram of a text processing device based on machine learning in an embodiment of the present application
  • Fig. 10 is a schematic diagram of a computer device in an embodiment of the present application.
  • the text processing method based on machine learning can be applied to the application environment shown in FIG. 1.
  • the text processing method based on machine learning is applied in a text processing system based on machine learning.
  • the text processing system based on machine learning includes a client and a server as shown in FIG. Communication is used to solve the problem of low accuracy of answers obtained by machine reading.
  • the client is also called the client, which refers to the program that corresponds to the server and provides local services to the client.
  • the client can be installed on, but not limited to, various personal computers, notebook computers, smart phones, tablet computers, and portable wearable devices.
  • the server can be implemented with a standalone server or a server cluster composed of multiple servers.
  • a text processing method based on machine learning is provided.
  • the application of the method to the server in FIG. 1 is taken as an example for description, including the following steps:
  • S10 Obtain the answer data to be processed, and preprocess the answer data to be processed to obtain standard answer data.
  • the standard answer data includes standard material information and standard question information.
  • the answer data to be processed refers to the reading comprehension data to be processed.
  • Each piece of reading comprehension data is regarded as a pending answer data.
  • the language of the answer data to be processed can be Chinese or English.
  • the answer data to be processed mainly includes reading materials and question information.
  • the topic information is mainly composed of questions and corresponding candidate answers.
  • the reading material can be single-paragraph text or multi-paragraph text.
  • a piece of reading material in the answer data to be processed may correspond to one or more item information.
  • any piece of reading comprehension data can be obtained directly from the test system, or any piece of reading comprehension data on the paper answer sheet can be scanned and recognized.
  • the preprocessing of the answer data to be processed mainly includes format judgment and processing of the answer data to be processed, to determine whether the format of the answer data to be processed meets preset conditions.
  • the answer data to be processed in English format can Input to the machine reading comprehension model for answer prediction. Therefore, if the text format of the answer data to be processed is Chinese, the answer data to be processed in Chinese format needs to be converted into the answer data to be processed in English format.
  • the answer data to be processed is assembled into the answer data in json format, and the json string in the answer data to be processed meets the requirements, such as judging the answer to be processed Whether the key in the data is vacant, whether the value type meets the requirements, whether the value length is within the range, etc. If the json string in the answer data to be processed does not meet the requirements, the answer data to be processed is returned to the client interface and Perform an abnormal display, prompting the user that the pending answer data is illegal data, and the pending answer data needs to be re-acquired.
  • the number of characters in the answer data to be processed exceeds the preset character threshold, then It is necessary to perform character segmentation processing on the answer data to be processed according to the real-time situation.
  • one answer data to be processed that originally contains a piece of reading material and multiple item information can be divided into multiple answer data to be processed, and each answer data to be processed includes One reading material and one topic information.
  • the standard answer data includes standard material information and standard question information.
  • the standard material information is material information that meets the requirements after preprocessing the material information in the answer data to be processed.
  • the standard item information is the item information that meets the requirements after preprocessing the item information in the answer data to be processed.
  • S20 Input the standard question information in the standard answer data into the preset answer classification model to obtain the question type of the standard question information.
  • one standard answer data may include one or more standard question information, and the question types corresponding to different standard question information may be different.
  • the standard question information included in a standard answer data may be a full-text inference question, a paragraph reasoning question, or a summary multiple-choice question.
  • the type of each standard question information in the standard answer data is determined.
  • the question type of each standard question information can be obtained.
  • the answer classification model is a pre-trained model that can identify the standard question information, thereby determining the question type of the standard question information.
  • the question type of the classified standard question information may be a vocabulary question, a highlight question, a full text inference question, an insertion question, a paragraph reasoning question, a summary multiple choice question, or a connection question.
  • the answer classification model is preferably a machine learning Bayesian model.
  • the machine learning Bayes model is obtained by training a large amount of topic information that has been classified and labeled in advance.
  • Bayesian decision theory Bayesian decision theory
  • Bayesian decision theory is the basic method to implement decision-making under the framework of probability. It is a combination of Decision theory + Probability theory. It discusses how to make optimal decisions in an environment containing uncertainty. For classification tasks, in an ideal situation where all relevant probabilities are known, Bayesian Decision theory considers how to select the optimal category label based on these probabilities and misjudgment losses (probability knowledge + knowledge of the loss caused by the decision ⁇ optimal decision).
  • S30 Input the standard material information, standard question information and the corresponding question type into the preset target machine reading comprehension model for prediction, and obtain initial answer information.
  • the initial answer information includes multiple evaluation data information and information corresponding to the standard question information.
  • Problem-solving ideas information where the target machine reading comprehension model is obtained by training using a convolutional neural network-pre-training language model.
  • the target machine reading comprehension model refers to a pre-trained model that can predict answers and analyze problem-solving ideas.
  • the target machine reading comprehension model is obtained by training with a convolutional neural network-pretrained language model.
  • Convolutional neural network-pre-training language model is a model obtained by combining convolutional neural network model and pre-training language model. Understandably, the convolutional neural network-pre-training language model is equivalent to the model formed by connecting the convolutional neural network and the pre-training language network model.
  • the target machine reading comprehension model mainly includes a prediction layer, a reasoning layer, an encoding layer, and a data layer.
  • the prediction layer includes several prediction units, and each prediction unit corresponds to one type of standard title information.
  • the prediction layer can include vocabulary item unit, highlight item unit, full-text inference item unit, insertion item unit, paragraph inference item unit, summary multiple-choice item unit, and connection item unit.
  • the reasoning layer mainly includes the RoBerta unit and the XLNet unit.
  • the RoBerta unit mainly obtains the selection probability value of each standard candidate text by combining the standard candidate text and the standard material information.
  • the XLNet unit mainly processes the standard candidate text and standard material information, and obtains the key information of the standard material information. Among them, the selection probability value is the probability value used to evaluate the standard candidate text as the correct answer.
  • the key information of the standard material information is the information after labeling and parsing each sentence in the standard material information. For example: mark which sentence in the standard material information is the central opinion sentence, the sub-thesis sentence, and the non-opinion sentence.
  • the coding layer is used to perform feature encoding on the selection probability value of each standard candidate text and the key information of the standard material information. And input the selection probability value of each standard candidate text for feature encoding and the key information of the standard material information into the data layer, so as to obtain the initial answer information.
  • the initial answer information includes multiple evaluation data information and problem-solving ideas information corresponding to the standard problem information.
  • the evaluation data information is the selection probability value corresponding to each candidate answer in the standard topic information. Since one standard question information includes at least two candidate answers, the initial answer information obtained includes multiple evaluation data information. Each candidate answer corresponds to an evaluation data message.
  • Problem-solving thinking information is the process of analyzing the normal answer derived from the standard topic information, that is, the reason and understanding process of why this answer was chosen.
  • S40 Determine final evaluation data from multiple evaluation data information according to the problem-solving idea information, and record the final evaluation data and the problem-solving idea information as target answer information in a preset integration manner.
  • each evaluation data information is a probability value corresponding to each candidate answer in the standard topic information. Therefore, after the probability value corresponding to each candidate answer in the standard question information is determined, the probability value corresponding to each candidate answer is screened according to the problem-solving idea information and the question requirements in the standard question information.
  • the final evaluation data is determined in the evaluation data information, that is, the correct answer corresponding to the standard question is determined, and then the final evaluation data corresponding to the standard question and the corresponding problem-solving idea information are recorded as the target answer in a preset integration method information.
  • the preset integration method can be to directly combine the final evaluation data and the corresponding problem-solving idea information.
  • the obtained initial answer information includes 4 evaluation data information, which are candidate answer A: 0.81, candidate answer B: 0.92, candidate answer C: 0.95 and candidate answer D: 0.01, the question requirements in the standard question information are Which is a conclusion that is impossible to infer from the material. Therefore, the final evaluation data is determined as the candidate answer D from the four evaluation data information combined with the problem-solving idea information. Understandably, the probability value corresponding to the candidate answer D is the smallest probability value, that is, the candidate answer D is unlikely to be inferred from the material, so the final evaluation data is the candidate answer D.
  • the final evaluation data and problem-solving ideas information are recorded as target answer information in a preset integration manner. Understandably, the target answer information includes the correct answer to the question and the reason and understanding process of why the answer was chosen.
  • the answer data to be processed is obtained, and the answer data to be processed is preprocessed to obtain standard answer data.
  • the standard answer data includes standard material information and standard question information; the standard question information in the standard answer data is input to the preset In the answer classification model, the question type of the standard question information is obtained; the standard material information, the standard question information and the corresponding question type are input into the preset target machine reading comprehension model for prediction, and the initial answer information is obtained.
  • the initial answer information includes Multiple evaluation data information and problem-solving ideas information corresponding to the standard topic information.
  • the target machine reading comprehension model is obtained by training using a convolutional neural network-pre-training language model; based on the problem-solving idea information from multiple evaluation data information Determine the final evaluation data in the process, and record the final evaluation data and the problem-solving idea information as the target answer information in a preset integration method; the target machine reading comprehension model obtained by using the convolutional neural network-pre-training language model training to deal with the answer
  • the data is used for answer prediction, and the target answer information that contains both the evaluation data information and the corresponding problem-solving idea information is obtained; thus, the accuracy and true meaning of the answers obtained by machine reading are further improved, which plays a role in assisting teaching/learning.
  • preprocessing the answer data to be processed includes the following steps:
  • S101 Standardize the text form of the answer data to be processed to obtain the initial answer data.
  • the language of the acquired answer data to be processed may be in Chinese format or English format
  • the answer data to be processed in English format can be input into the machine reading comprehension model for answer prediction, therefore,
  • the text format of the answer data to be processed is standardized, that is, the answer data to be processed is converted into a unified English format to obtain the initial answer data.
  • the initial answer data is assembled into candidate answer data in json format.
  • the json data format is a lightweight data exchange format that uses a text format completely independent of programming languages to store and represent data.
  • the concise and clear hierarchical structure of the json data format is not only easy for humans to read and write, but also easy for machine analysis and generation, and can effectively improve network transmission efficiency. Therefore, by converting the initial answer data into the json data format, it is beneficial to the subsequent rapid and accurate data processing.
  • classes or functions that convert various data formats (map, xml or yaml, etc.) into json data format can be pre-written and packaged into a conversion script to convert the initial answer data into candidate answer data in json data format.
  • When performing data format conversion first obtain the corresponding conversion scripts according to the data format of the initial answer data, and then execute the corresponding conversion scripts to convert the initial answer data into a json data format to obtain candidate answer data.
  • S103 Determine whether the json character string in the candidate answer data meets the preset requirement, and if the json character string in the candidate answer data meets the preset requirement, determine the candidate answer data as the standard answer data.
  • judging whether the json string in the candidate answer data meets the preset requirements is mainly to judge whether the key in the json string is vacant, whether the value type meets the requirements, and whether the value length is within the range, etc.
  • the preset type range and the preset length range of the value in the json string that meet the requirements have been preset. If the key in the json string in the candidate answer data is not empty, the value type is within the preset type range, and the length of the value is within the preset length range, it is determined that the json string in the candidate answer data meets the preset requirements , Determine the candidate answer data as the standard answer data.
  • the answer data to be processed is returned to the client interface and an abnormal display is performed, prompting the user that the answer data to be processed is illegal data, and the answer data to be processed needs to be retrieved.
  • the text format of the answer data to be processed is standardized to obtain the initial answer data; the initial answer data is converted into a json data format to obtain candidate answer data; it is judged whether the json string in the candidate answer data meets the preset requirements , If the json string in the candidate answer data meets the preset requirements, the candidate answer data is determined as the standard answer data; thereby improving the accuracy and uniformity of the obtained standard answer data, and ensuring that the subsequent data is input to the target machine The accuracy of the predictions made in the reading comprehension model.
  • inputting standard material information, standard question information and corresponding question types into a preset target machine reading comprehension model for prediction to obtain initial answer information specifically includes the following steps:
  • S301 Input standard material information, standard topic information, and corresponding topic types into the prediction layer of the target machine's reading comprehension model to obtain a standard candidate text set of standard topic information.
  • the standard candidate text set includes at least one standard preparation Select the text.
  • the standard candidate text set refers to the text set obtained by separately concatenating the question in the standard topic information and each candidate answer.
  • the standard candidate text set contains at least one standard candidate text.
  • the standard material information, the standard topic information, and the corresponding topic type are input into the prediction layer of the target machine reading comprehension model.
  • the processing logic of the prediction layer corresponding to different types of standard title information is different. That is, multiple types of processing units are included in the prediction layer of the target machine reading comprehension model.
  • the prediction layer of the target machine reading comprehension model includes a vocabulary item unit, a highlight item unit, a full-text inference item unit, an insertion item unit, a paragraph inference item unit, a summary multiple-choice item unit, and a connection item. unit.
  • the topic type of the standard topic information is a vocabulary question
  • the standard topic information is input into the prediction layer of the target machine's reading comprehension model, it will be based on the topic type associated with the standard topic information: vocabulary question.
  • the standard topic information is automatically input into the vocabulary item unit of the prediction layer of the target machine's reading comprehension model, so as to obtain the standard candidate text set of the standard topic information.
  • S302 Input each standard candidate text and standard material information in the standard candidate text set into the inference layer of the target machine reading comprehension model to obtain the selection probability value of each standard candidate text and key information of the standard material information.
  • the reasoning layer is used to judge whether each standard candidate text can be inferred from the standard material information.
  • the reasoning layer includes RoBerta unit and XLNet unit.
  • RoBERTa is the enhancement and tuning of BERT.
  • RoBERTa mainly made improvements to the previously proposed BERT in three aspects. One is the specific details of the model and improved the optimization function; the second is the training strategy level, which uses a dynamic mask to train the model, which proves the NSP (Next Sentence Prediction) The lack of training strategy uses a larger batch size; the third is the data level, on the one hand, a larger data set is used, on the other hand, BPE (Byte-Pair Encoding) is used to process text data .
  • XLNet is a general autoregressive pre-training method that learns bidirectional contextual information by maximizing the log likelihood of all possible factorization orders.
  • each standard candidate text and standard material information in the standard candidate text set output by the prediction layer is input into the inference layer of the target machine reading comprehension model; the RoBerta unit is used to process the standard candidate text and standard material information , So as to obtain the selection probability value of each standard candidate text, and use the XLNet unit to process the standard candidate text and the standard material information to obtain the key information of the standard material information.
  • the selection probability value is the probability value used to evaluate the standard candidate text as the correct answer.
  • the range of the selection probability value is 0-1. The higher the selection probability value, the greater the probability that the corresponding standard candidate text is the correct answer.
  • the key information of the standard material information is the information after labeling and parsing each sentence in the standard material information. For example: which of the standard material information are the central opinion sentence, the sub-thesis sentence and the non-opinion sentence, etc.
  • the target machine reading comprehension model also includes an encoding layer and a data layer.
  • the encoding layer is mainly responsible for feature encoding of the standard candidate text and standard material information input to the inference layer.
  • the encoding layer mainly uses the BERT encoder method and the XLNet encoder method standard candidate text and standard material information for feature encoding.
  • Feature coding The problem that the data layer solves is the dependence of the Base model, because our reasoning model is not from 0 to 1, but is based on the industry's large-scale training model to do some migration, so the data we are based on include RACE, SQuAD, etc.
  • S303 Combine the selection probability value of each standard candidate text and the key information of the standard material information to obtain initial answer information.
  • the selection probability value of each standard candidate text and the key information of the standard material information are combined, namely The initial answer information can be obtained.
  • standard material information, standard topic information, and corresponding topic types are input into the prediction layer of the target machine's reading comprehension model to obtain a standard candidate text set of standard topic information.
  • the standard candidate text set includes At least one standard candidate text; input each standard candidate text and standard material information in the standard candidate text set into the inference layer of the target machine reading comprehension model to obtain the selection probability value and standard material of each standard candidate text.
  • the key information of the information; the selection probability value of each standard candidate text and the key information of the standard material information are combined to obtain the initial answer information; thereby improving the accuracy of the generated initial answer information.
  • the machine learning-based text processing method before the standard material information, standard topic information, and corresponding topic types are input into a preset target machine reading comprehension model for prediction, the machine learning-based text processing method also Specifically include the following steps:
  • S11 Obtain a preset number of sample answer data, and each sample answer data includes key paragraph information, sample questions, and corresponding candidate answer sets.
  • the sample answer data refers to the reading comprehension data used for model training.
  • the sample answer data can be obtained by directly acquiring several pieces of reading comprehension data from the test system, or by scanning and identifying the reading comprehension data on the paper answer sheet.
  • Each of the sample answer data includes key paragraph information, sample questions, and corresponding candidate answer sets.
  • the key paragraph information refers to the material information corresponding to the sample question.
  • the sample question refers to the question of the question in the sample answer data.
  • the sample question and the corresponding candidate answer set are the candidate answer items corresponding to the sample question.
  • the preset number can be M, where M is a positive integer.
  • M is a positive integer.
  • the specific value of M can be set according to actual needs. The higher the value of M, the higher the accuracy of subsequent model training, but the extraction efficiency will decrease, and the selection of M can be comprehensively considered in terms of accuracy and efficiency.
  • S12 Combine the sample question of each sample answer data with each candidate answer in the corresponding candidate answer set to obtain a sample candidate text set of each sample answer data.
  • the sample candidate text set includes at least one sample Alternative text.
  • sample questions of each sample answer data and each candidate answer in the corresponding candidate answer set are respectively spliced to obtain at least one sample candidate text of each sample answer data.
  • the key paragraph information of each sample answer data is annotated to obtain annotation data of the key paragraph information
  • annotation data is data used to annotate the key information of each sentence in the key paragraph information.
  • the labeling data can be used to label which sentences in the key paragraph information are central point sentences, which sentences are sub-thesis sentences, and which sentences are non-point sentences.
  • S14 Input the sample candidate text set, key paragraph information, and corresponding annotation data in each sample answer data as training samples into the convolutional neural network-pre-training language model for training, to obtain the target machine reading comprehension model.
  • the sample candidate text set, key paragraph information, and corresponding annotation data in each sample answer data are input as training samples into the convolutional neural network-pre-training language model for training, and the target machine reading comprehension can be obtained Model.
  • the convolutional neural network-pre-training language model is a model obtained by combining the convolutional neural network model and the pre-training language model. Understandably, the convolutional neural network-pre-training language model is equivalent to the model formed by connecting the convolutional neural network and the pre-training language network model.
  • a preset number of sample answer data is obtained, and each sample answer data includes key paragraph information, sample questions, and corresponding candidate answer sets; the sample questions of each sample answer data and the corresponding candidate
  • Each candidate answer in the answer set is spliced to obtain a sample candidate text set of each sample answer data.
  • the sample candidate text set includes at least one sample candidate text; the key paragraph information of each sample answer data is marked, Obtain the annotation data of key paragraph information; input the sample candidate text set, key paragraph information and corresponding annotation data in each sample answer data as training samples into the convolutional neural network-pre-training language model for training, and obtain the target Machine reading comprehension model; thereby improving the accuracy of the generated target machine reading comprehension model.
  • the sample candidate text set, key paragraph information, and corresponding annotation data in each sample answer data are input as training samples into the convolutional neural network-pre-training language model
  • the text comprehension processing method based on machine learning also specifically includes the following steps:
  • S15 Receive an update instruction, and detect whether the minimum risk training loss function in the target machine's reading comprehension model is minimized.
  • the update instruction refers to an instruction used to trigger the optimization of the target machine's reading comprehension model.
  • the update instruction may be generated when the target machine's reading comprehension model is required to have a more accurate predictive ability, or a trigger cycle may be preset for periodic generation, etc. Specifically, an update instruction is received, and it is detected whether the minimum risk training loss function in the reading comprehension model of the target machine is minimized.
  • the goal is to minimize the minimum risk training loss function, and the parameters of the target machine reading comprehension model are optimized for a preset number of times, and then the target is executed
  • the training of the machine reading comprehension model continuously optimizes the probability distribution of the output answers of the target machine reading comprehension model, so that the answers to the sample questions in the predicted sample answer data are getting closer and closer to the standard answers. Therefore, through a preset number of iterative optimization adjustments, an adjusted target machine reading comprehension model can be obtained.
  • the minimum risk training refers to the use of the loss function ⁇ (y,y (n) ) to describe the degree of difference between the answer y predicted by the model and the standard answer y (n) , and to try to find a set of parameters to make the model in the training set The expected value of the loss.
  • x (n) is the sample question in the sample answer data
  • y is the answer output by the target machine reading comprehension model
  • ⁇ ) is the target machine reading comprehension model when the model parameter is ⁇
  • Y(x (n) ) is the set of all possible output answers of the target machine reading comprehension model corresponding to x (n)
  • ⁇ (y,y (n) ) is the answer output by the target machine reading comprehension model
  • the rouge evaluation in this embodiment adopts rouge-L, and the corresponding calculation formula is: in the above formula, x and y are the text sequence of the standard answer and the model output answer; N is The length of the standard answer; n is the length of the model output answer; ⁇ is a hyperparameter, which can be set as required, and the value is 1.2 in this embodiment; LCS is the longest common subsequence.
  • the preset evaluation function and selected verification answer data are used to evaluate the accuracy of the adjusted target machine reading comprehension model output answer. Obtain the evaluation result; among them, an optimization adjustment is performed on the parameters of the target machine reading comprehension model, including a minimization process for the minimum risk training loss function.
  • the evaluation result refers to the result obtained after the effect evaluation of the target machine reading comprehension model after parameter adjustment.
  • the verification answer data refers to the data set used to verify the effect of the target machine's reading comprehension model after parameter adjustment.
  • Each verification answer data includes key paragraph information, sample questions and corresponding candidate answer sets.
  • the target machine reading comprehension model is optimized and adjusted for a preset number of times, the selected verification answer data is input into the adjusted target machine reading comprehension model, and then the preset evaluation function is used, such as ROUGE (Recall-Oriented Understudy ForGisting Evaluation, evaluation of the understanding of improvement evaluation, BLEU (Bilingual Evaluation Understudy, bilingual evaluation) evaluates the accuracy of the answers output by the adjusted target machine reading comprehension model, and obtains the evaluation result.
  • ROUGE Recall-Oriented Understudy ForGisting Evaluation, evaluation of the understanding of improvement evaluation
  • BLEU Bilingual Evaluation Understudy, bilingual evaluation
  • the evaluation result it is determined whether the evaluation result meets the preset evaluation requirements. If the evaluation result meets the preset evaluation requirements, the optimization adjustment of the target machine reading comprehension model is stopped, and the adjusted target machine reading comprehension The model is recorded as a new target machine reading comprehension model.
  • the preset evaluation requirement is when the loss function in the reading comprehension model of the target machine reaches the minimum until it converges. That is, when the evaluation result indicates that the loss function in the target machine reading comprehension model converges during the iterative optimization and adjustment process, and the minimum optimized loss function is obtained, it means that the evaluation result meets the preset evaluation requirements, and the optimization of the target machine reading comprehension model is stopped.
  • the obtained evaluation result does not meet the preset evaluation requirements, continue to optimize and adjust the target machine reading comprehension model to minimize the loss function until it converges, until the evaluation result meets the preset Assess the requirements, and finally record the adjusted target machine reading comprehension model as the new target machine reading comprehension model.
  • the target machine reading comprehension model performs an iterative optimization adjustment, an evaluation result will be output accordingly, so that after a preset number of iterative optimization adjustments and evaluations, multiple evaluations will be correspondingly obtained.
  • an update instruction is received to detect whether the minimum risk training loss function in the target machine reading comprehension model is minimized; when the minimum risk training loss function is not minimized, the parameters of the target machine reading comprehension model are preset After the optimization and adjustment of the number of times, the preset evaluation function and the selected verification answer data are used to evaluate the accuracy of the adjusted target machine reading comprehension model output answer, and the evaluation result is obtained; among them, the parameters of the target machine reading comprehension model are evaluated.
  • An optimization adjustment including a minimization process for the minimum risk training loss function; if the evaluation result meets the preset evaluation requirements, the adjusted target machine reading comprehension model is recorded as a new target machine reading comprehension model, so as to facilitate Standard material information, standard question information and corresponding question types are re-input into the new target machine reading comprehension model for prediction, and initial answer information is obtained, thereby further improving the accuracy and accuracy of the obtained initial answer information.
  • a text processing device based on machine learning is provided, and the text processing device based on machine learning has a one-to-one correspondence with the text processing method based on machine learning in the foregoing embodiment.
  • the machine learning-based text processing device includes a preprocessing module, a first input module 20, a prediction module 30, and an integration module 40.
  • the detailed description of each functional module is as follows:
  • the preprocessing module 10 is configured to obtain answer data to be processed, and preprocess the answer data to be processed to obtain standard answer data, where the standard answer data includes standard material information and standard question information;
  • the first input module 20 is configured to input the standard question information in the standard answer data into a preset answer classification model to obtain the question type of the standard question information;
  • the prediction module 30 is configured to input the standard material information, the standard question information, and the corresponding question type into a preset target machine reading comprehension model for prediction to obtain initial answer information, where the initial answer information includes A plurality of evaluation data information and problem-solving ideas information corresponding to the standard topic information, wherein the target machine reading comprehension model is obtained by training using a convolutional neural network-pre-training language model;
  • the determining module 40 is configured to determine final evaluation data from a plurality of the evaluation data information according to the problem-solving idea information, and record the final evaluation data and the problem-solving idea information in a preset integration manner as a target Answer information.
  • the preprocessing module 10 includes:
  • the standardization unit 101 is used to standardize the text form of the answer data to be processed to obtain initial answer data;
  • the conversion unit 102 is configured to convert the initial answer data into a json data format to obtain candidate answer data
  • the judging unit 103 is configured to judge whether the json character string in the candidate answer data meets preset requirements, and if the json character string in the candidate answer data meets the preset requirements, determine the candidate answer data as a standard answer data.
  • the prediction module 30 includes:
  • the first input unit 301 is configured to input the standard material information, the standard topic information, and the corresponding topic type into the prediction layer of the target machine reading comprehension model to obtain the standard topic information A selected text set, where the standard candidate text set includes at least one standard candidate text;
  • the second input unit 302 is configured to input each standard candidate text and the standard material information in the standard candidate text set into the inference layer of the target machine reading comprehension model to obtain each The selection probability value of the standard candidate text and the key information of the standard material information;
  • the combining unit 303 is configured to combine the selection probability value of each standard candidate text and the key information of the standard material information to obtain initial answer information.
  • the text processing device based on machine learning further includes:
  • the obtaining module is used to obtain a preset number of sample answer data, each of the sample answer data includes key paragraph information, sample questions, and corresponding candidate answer sets;
  • the splicing module is used to splice the sample question of each sample answer data with each candidate answer in the corresponding candidate answer set to obtain the sample candidate text of each sample answer data Set, the sample candidate text set includes at least one sample candidate text;
  • An annotation module configured to annotate the key paragraph information of each of the sample answer data to obtain the annotation data of the key paragraph information
  • the second input module is used to input the sample candidate text set, the key paragraph information and the corresponding annotation data in each of the sample answer data as training samples into the convolutional neural network-pre-training language model Perform training to obtain the target machine reading comprehension model.
  • the text processing device based on machine learning further includes:
  • the detection module is configured to receive update instructions and detect whether the minimum risk training loss function in the target machine reading comprehension model is minimized;
  • the optimization adjustment module is used to optimize and adjust the parameters of the target machine reading comprehension model for a preset number of times when the minimum risk training loss function is not minimized, and then use the preset evaluation function and the selected verification answer data, Evaluate the accuracy of the output answers of the adjusted target machine reading comprehension model to obtain the evaluation result; wherein, performing an optimization adjustment on the parameters of the target machine reading comprehension model includes performing the minimum risk training loss function Minimize the processing flow at one time;
  • the recording module is used to record the adjusted target machine reading comprehension model as a new target machine reading comprehension model when the evaluation result meets the preset evaluation requirements, so as to facilitate the standard material information, standard topic information and corresponding
  • the question type of is re-input into the new target machine reading comprehension model for prediction, and the initial answer information is obtained.
  • the various modules in the above-mentioned machine learning-based text processing device can be implemented in whole or in part by software, hardware, and a combination thereof.
  • the foregoing modules may be embedded in the form of hardware or independent of the processor in the computer device, or may be stored in the memory of the computer device in the form of software, so that the processor can call and execute the operations corresponding to the foregoing modules.
  • a computer device is provided.
  • the computer device may be a server, and its internal structure diagram may be as shown in FIG. 10.
  • the computer equipment includes a processor, a memory, a network interface, and a database connected through a system bus.
  • the processor of the computer device is used to provide calculation and control capabilities.
  • the memory of the computer device includes a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium stores an operating system, a computer program, and a database.
  • the internal memory provides an environment for the operation of the operating system and computer programs in the non-volatile storage medium.
  • the database of the computer device is used to store the data used in the text processing method based on machine learning in the foregoing embodiment.
  • the network interface of the computer device is used to communicate with an external terminal through a network connection.
  • the computer program is executed by the processor to realize a text processing method based on machine learning.
  • a computer device including a memory, a processor, and a computer program stored in the memory and capable of running on the processor.
  • the processor executes the computer program, it implements the machine learning-based Text processing method.
  • a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, the machine learning-based text processing method in the foregoing embodiment is implemented.
  • the computer-readable storage medium may be non-volatile or volatile.
  • Non-volatile memory may include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable programmable ROM (EEPROM), or flash memory.
  • Volatile memory may include random access memory (RAM) or external cache memory.
  • RAM is available in many forms, such as static RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDRSDRAM), enhanced SDRAM (ESDRAM), synchronous chain Channel (Synchlink) DRAM (SLDRAM), memory bus (Rambus) direct RAM (RDRAM), direct memory bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM), etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computational Linguistics (AREA)
  • Evolutionary Computation (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Databases & Information Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Software Systems (AREA)
  • Computing Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Human Computer Interaction (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Machine Translation (AREA)

Abstract

Sont divulgués un procédé et un appareil de traitement de texte fondés sur un apprentissage automatique ainsi qu'un dispositif informatique et un support, qui se rapportent au domaine de la prise intelligente de décisions en intelligence artificielle. Le procédé consiste : à acquérir des données de réponse à une question à traiter, et à prétraiter lesdites données de réponse à une question afin d'obtenir des données de réponse à une question standard, les données de réponse à une question standard comprenant des informations pertinentes standard et des informations de question standard (S10) ; à entrer les informations de question standard dans les données de réponse à une question standard dans un modèle de classification de réponses à une question prédéfini afin d'obtenir un type de question des informations de question standard (S20) ; à entrer les informations pertinentes standard, les informations de question standard et le type de question correspondant dans un modèle de compréhension de lecture automatique cible prédéfini pour une prédiction afin d'obtenir des informations de réponse initiale, les informations de réponse initiale comprenant une pluralité d'éléments d'informations de données d'évaluation et d'informations de pensées de résolution de problème correspondant aux informations de question standard, le modèle de compréhension de lecture automatique cible étant obtenu au moyen d'un entraînement à l'aide d'un modèle de langage de pré-entraînement de réseau neuronal à convolution (S30) ; et à déterminer des données d'évaluation finale parmi la pluralité d'éléments d'informations de données d'évaluation selon les informations de pensée de résolution de problème, et à enregistrer les données d'estimation finale et les informations de pensée de résolution de problème en tant qu'informations de réponse cible dans un mode d'intégration prédéfini (S40). Le procédé améliore la précision d'une réponse obtenue au moyen d'une lecture automatique.
PCT/CN2020/103784 2020-06-05 2020-07-23 Procédé et appareil de traitement de texte fondés sur un apprentissage automatique et dispositif informatique et support WO2021243828A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010502599.4 2020-06-05
CN202010502599.4A CN111428021B (zh) 2020-06-05 2020-06-05 基于机器学习的文本处理方法、装置、计算机设备及介质

Publications (1)

Publication Number Publication Date
WO2021243828A1 true WO2021243828A1 (fr) 2021-12-09

Family

ID=71555335

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/103784 WO2021243828A1 (fr) 2020-06-05 2020-07-23 Procédé et appareil de traitement de texte fondés sur un apprentissage automatique et dispositif informatique et support

Country Status (2)

Country Link
CN (1) CN111428021B (fr)
WO (1) WO2021243828A1 (fr)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114358425A (zh) * 2022-01-06 2022-04-15 杭州数理大数据技术有限公司 一种试题选择方法、装置、电子设备及存储介质
CN114462355A (zh) * 2022-01-17 2022-05-10 网易有道信息技术(北京)有限公司 题目获取方法、装置、电子设备及存储介质
CN114707516A (zh) * 2022-03-29 2022-07-05 北京理工大学 一种基于对比学习的长文本语义相似度计算方法
CN114861597A (zh) * 2022-05-17 2022-08-05 北京飞象星球科技有限公司 填空题解题模型的训练方法及装置
CN114971956A (zh) * 2022-04-15 2022-08-30 北京飞象星球科技有限公司 题目解答方法及装置
CN115434870A (zh) * 2022-07-29 2022-12-06 国家电投集团云南国际电力投资有限公司 风力发电组振动数据的处理方法、装置、设备及存储介质
CN115801888A (zh) * 2022-11-16 2023-03-14 云南电网有限责任公司 一种基于通信技术的安全无线通信***
CN115862031A (zh) * 2022-12-30 2023-03-28 北京百度网讯科技有限公司 文本处理方法、神经网络的训练方法、装置和设备
CN116361463A (zh) * 2023-03-27 2023-06-30 应急管理部国家减灾中心(应急管理部卫星减灾应用中心) 一种地震灾情信息提取方法、装置、设备及介质
CN116484091A (zh) * 2023-03-10 2023-07-25 湖北天勤伟业企业管理有限公司 卡牌信息程序交互方法和装置
CN116798036A (zh) * 2023-06-27 2023-09-22 广州市南方人力资源评价中心有限公司 用于识别和校验答题卡客观题识别结果的方法及装置
CN117252739A (zh) * 2023-11-17 2023-12-19 山东山大鸥玛软件股份有限公司 一种评卷方法、***、电子设备及存储介质
CN117786416A (zh) * 2024-02-28 2024-03-29 腾讯科技(深圳)有限公司 一种模型训练方法、装置、设备、存储介质及产品
CN118093831A (zh) * 2024-04-15 2024-05-28 清华大学 文本评测基准构建方法及装置
CN118114772A (zh) * 2024-04-30 2024-05-31 山东蓝思教育科技(集团)有限公司 一种题库答案解析方法及***

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111428021B (zh) * 2020-06-05 2023-05-30 平安国际智慧城市科技股份有限公司 基于机器学习的文本处理方法、装置、计算机设备及介质
CN112183048A (zh) * 2020-09-30 2021-01-05 北京有竹居网络技术有限公司 一种自动解题方法、装置、计算机设备及存储介质
CN112560431A (zh) * 2020-12-25 2021-03-26 北京百度网讯科技有限公司 用于生成试题辅导信息的方法、装置、设备、存储介质以及计算机程序产品
CN112764760B (zh) * 2021-01-25 2023-12-26 中国科学院自动化研究所 基于程序评测的辅助答题***
CN113220847B (zh) * 2021-06-01 2024-06-21 平安科技(深圳)有限公司 基于神经网络的知识掌握程度测评方法、装置及相关设备
CN113407675A (zh) * 2021-06-24 2021-09-17 作业帮教育科技(北京)有限公司 教育题目自动批改方法、装置和电子设备
CN113468176B (zh) * 2021-06-30 2023-09-26 北京百度网讯科技有限公司 信息录入方法和装置、电子设备、计算机可读存储介质
CN113488178B (zh) * 2021-07-20 2022-07-12 上海弗莱特智能医疗科技有限公司 一种信息生成方法、装置、存储介质及电子设备
CN113779360A (zh) * 2021-08-18 2021-12-10 深圳技术大学 基于多头问答模型的解题方法、装置、设备及存储介质
CN113805695B (zh) * 2021-08-26 2024-04-05 深圳静美大健康科技有限公司 阅读理解水平的预测方法及装置、电子设备和存储介质
CN114330718B (zh) * 2021-12-23 2023-03-24 北京百度网讯科技有限公司 因果关系的提取方法、装置及电子设备
CN114444488B (zh) * 2022-01-26 2023-03-24 中国科学技术大学 一种少样本机器阅读理解方法、***、设备及存储介质
CN117892818B (zh) * 2024-03-18 2024-05-28 浙江大学 一种基于隐式思维链的大语言模型推理性内容生成方法

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110619123A (zh) * 2019-09-19 2019-12-27 电子科技大学 一种机器阅读理解方法
CN110688491A (zh) * 2019-09-25 2020-01-14 暨南大学 基于深度学习的机器阅读理解方法、***、设备及介质
CN110750630A (zh) * 2019-09-25 2020-02-04 北京捷通华声科技股份有限公司 一种生成式机器阅读理解方法、装置、设备及存储介质
US20200134449A1 (en) * 2018-10-26 2020-04-30 Naver Corporation Training of machine reading and comprehension systems
CN111428021A (zh) * 2020-06-05 2020-07-17 平安国际智慧城市科技股份有限公司 基于机器学习的文本处理方法、装置、计算机设备及介质

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9720981B1 (en) * 2016-02-25 2017-08-01 International Business Machines Corporation Multiple instance machine learning for question answering systems
CN108229718B (zh) * 2016-12-22 2020-06-02 北京字节跳动网络技术有限公司 一种信息预测方法及装置
CN111027327B (zh) * 2019-10-29 2022-09-06 平安科技(深圳)有限公司 机器阅读理解方法、设备、存储介质及装置

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200134449A1 (en) * 2018-10-26 2020-04-30 Naver Corporation Training of machine reading and comprehension systems
CN110619123A (zh) * 2019-09-19 2019-12-27 电子科技大学 一种机器阅读理解方法
CN110688491A (zh) * 2019-09-25 2020-01-14 暨南大学 基于深度学习的机器阅读理解方法、***、设备及介质
CN110750630A (zh) * 2019-09-25 2020-02-04 北京捷通华声科技股份有限公司 一种生成式机器阅读理解方法、装置、设备及存储介质
CN111428021A (zh) * 2020-06-05 2020-07-17 平安国际智慧城市科技股份有限公司 基于机器学习的文本处理方法、装置、计算机设备及介质

Cited By (22)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114358425A (zh) * 2022-01-06 2022-04-15 杭州数理大数据技术有限公司 一种试题选择方法、装置、电子设备及存储介质
CN114462355A (zh) * 2022-01-17 2022-05-10 网易有道信息技术(北京)有限公司 题目获取方法、装置、电子设备及存储介质
CN114707516A (zh) * 2022-03-29 2022-07-05 北京理工大学 一种基于对比学习的长文本语义相似度计算方法
CN114971956A (zh) * 2022-04-15 2022-08-30 北京飞象星球科技有限公司 题目解答方法及装置
CN114861597A (zh) * 2022-05-17 2022-08-05 北京飞象星球科技有限公司 填空题解题模型的训练方法及装置
CN115434870A (zh) * 2022-07-29 2022-12-06 国家电投集团云南国际电力投资有限公司 风力发电组振动数据的处理方法、装置、设备及存储介质
CN115434870B (zh) * 2022-07-29 2023-11-07 云南滇能智慧能源有限公司 风力发电组振动数据的处理方法、装置、设备及存储介质
CN115801888A (zh) * 2022-11-16 2023-03-14 云南电网有限责任公司 一种基于通信技术的安全无线通信***
CN115862031A (zh) * 2022-12-30 2023-03-28 北京百度网讯科技有限公司 文本处理方法、神经网络的训练方法、装置和设备
CN115862031B (zh) * 2022-12-30 2024-02-20 北京百度网讯科技有限公司 文本处理方法、神经网络的训练方法、装置和设备
CN116484091A (zh) * 2023-03-10 2023-07-25 湖北天勤伟业企业管理有限公司 卡牌信息程序交互方法和装置
CN116484091B (zh) * 2023-03-10 2024-07-19 湖北天勤伟业企业管理有限公司 卡牌信息程序交互方法和装置
CN116361463A (zh) * 2023-03-27 2023-06-30 应急管理部国家减灾中心(应急管理部卫星减灾应用中心) 一种地震灾情信息提取方法、装置、设备及介质
CN116361463B (zh) * 2023-03-27 2023-12-08 应急管理部国家减灾中心(应急管理部卫星减灾应用中心) 一种地震灾情信息提取方法、装置、设备及介质
CN116798036A (zh) * 2023-06-27 2023-09-22 广州市南方人力资源评价中心有限公司 用于识别和校验答题卡客观题识别结果的方法及装置
CN116798036B (zh) * 2023-06-27 2024-04-02 网才科技(广州)集团股份有限公司 用于识别和校验答题卡客观题识别结果的方法及装置
CN117252739B (zh) * 2023-11-17 2024-03-12 山东山大鸥玛软件股份有限公司 一种评卷方法、***、电子设备及存储介质
CN117252739A (zh) * 2023-11-17 2023-12-19 山东山大鸥玛软件股份有限公司 一种评卷方法、***、电子设备及存储介质
CN117786416A (zh) * 2024-02-28 2024-03-29 腾讯科技(深圳)有限公司 一种模型训练方法、装置、设备、存储介质及产品
CN117786416B (zh) * 2024-02-28 2024-05-28 腾讯科技(深圳)有限公司 一种模型训练方法、装置、设备、存储介质及产品
CN118093831A (zh) * 2024-04-15 2024-05-28 清华大学 文本评测基准构建方法及装置
CN118114772A (zh) * 2024-04-30 2024-05-31 山东蓝思教育科技(集团)有限公司 一种题库答案解析方法及***

Also Published As

Publication number Publication date
CN111428021B (zh) 2023-05-30
CN111428021A (zh) 2020-07-17

Similar Documents

Publication Publication Date Title
WO2021243828A1 (fr) Procédé et appareil de traitement de texte fondés sur un apprentissage automatique et dispositif informatique et support
US12039280B2 (en) Multi-turn dialogue response generation with persona modeling
US11948058B2 (en) Utilizing recurrent neural networks to recognize and extract open intent from text inputs
CN110717039A (zh) 文本分类方法和装置、电子设备、计算机可读存储介质
US10970493B1 (en) Systems and methods for slot relation extraction for machine learning task-oriented dialogue systems
CN112287089B (zh) 用于自动问答***的分类模型训练、自动问答方法及装置
US20220351634A1 (en) Question answering systems
EP3832485A1 (fr) Systèmes de réponse à des questions
CN113672931B (zh) 一种基于预训练的软件漏洞自动检测方法及装置
US20220067277A1 (en) Intelligent Training Set Augmentation for Natural Language Processing Tasks
CN110377733A (zh) 一种基于文本的情绪识别方法、终端设备及介质
CN115391527A (zh) 基于提示学习的意图识别方法、问答方法及装置
CN112101042A (zh) 文本情绪识别方法、装置、终端设备和存储介质
WO2021174814A1 (fr) Procédé et appareil de vérification de réponses pour une tâche d'externalisation ouverte, dispositif informatique et support d'informations
CN113505786A (zh) 试题拍照评判方法、装置及电子设备
CN117634490A (zh) 信息抽取方法、装置、设备、介质和程序产品
CN117520497A (zh) 大模型交互处理方法、***、终端、设备及介质
Huang et al. Multitask learning for neural generative question answering
CN115374278A (zh) 文本处理模型蒸馏方法、装置、计算机设备及介质
US11562137B2 (en) System to correct model drift for natural language understanding
WO2023173554A1 (fr) Procédé et appareil d'identification de langage d'agent inapproprié, dispositif électronique et support de stockage
Mihaylov et al. Predicting the Resolution Time and Priority of Bug Reports: A Deep Learning Approach
US20220366197A1 (en) System and method for finetuning automated sentiment analysis
US20240062219A1 (en) Granular taxonomy for customer support augmented with ai
CN117725198A (zh) 搜索结果的摘要纠正方法、装置、设备及存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20938739

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 17.03.2023)

122 Ep: pct application non-entry in european phase

Ref document number: 20938739

Country of ref document: EP

Kind code of ref document: A1