CN112528662A - Entity category identification method, device, equipment and storage medium based on meta-learning - Google Patents

Entity category identification method, device, equipment and storage medium based on meta-learning Download PDF

Info

Publication number
CN112528662A
CN112528662A CN202011472865.XA CN202011472865A CN112528662A CN 112528662 A CN112528662 A CN 112528662A CN 202011472865 A CN202011472865 A CN 202011472865A CN 112528662 A CN112528662 A CN 112528662A
Authority
CN
China
Prior art keywords
sample
data
entity
query
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011472865.XA
Other languages
Chinese (zh)
Inventor
刘玉
徐国强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
OneConnect Smart Technology Co Ltd
OneConnect Financial Technology Co Ltd Shanghai
Original Assignee
OneConnect Financial Technology Co Ltd Shanghai
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by OneConnect Financial Technology Co Ltd Shanghai filed Critical OneConnect Financial Technology Co Ltd Shanghai
Priority to CN202011472865.XA priority Critical patent/CN112528662A/en
Publication of CN112528662A publication Critical patent/CN112528662A/en
Priority to PCT/CN2021/109617 priority patent/WO2022127124A1/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/289Phrasal analysis, e.g. finite state techniques or chunking
    • G06F40/295Named entity recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities
    • G06F40/284Lexical analysis, e.g. tokenisation or collocates
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Software Systems (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Medical Informatics (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

The present application relates to the field of artificial intelligence, and in particular, to a method, an apparatus, a device, and a storage medium for entity category identification based on meta-learning. The method comprises the following steps: acquiring a newly added entity type, and inquiring a reference sample corresponding to the newly added entity type; acquiring data to be identified; and inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model so as to recognize a new entity class corresponding to the reference sample in each data to be recognized, wherein the entity class recognition model is obtained by training based on a meta-learning mode. By adopting the method, the entity category identification accuracy can be ensured. In addition, the application also relates to a block chain technology, and the newly added entity type can be stored in the block chain node.

Description

Entity category identification method, device, equipment and storage medium based on meta-learning
Technical Field
The present application relates to the field of artificial intelligence technologies, and in particular, to a method, an apparatus, a device, and a storage medium for entity category identification based on meta-learning.
Background
At present, the research on named entity recognition in the field of artificial intelligence is more, but data sets related to named entity recognition are few, and particularly, Chinese named entity recognition data sets are rare. In addition, although a set of relatively mature named entity recognition models exist in the market, the models can only distinguish three relatively common entity categories, namely, names, organizations and addresses. These models cannot handle when new entity classes appear.
The traditional open-source Chinese named entity identification data set mainly comprises MSRA (Chinese Mobile architecture), civil daily report, microblog, CLUENER (customer premise equipment) and BOSON (business oriented network) data sets, the size of the five data sets is not very large, the number of entity categories is not large enough, and the total number of the five data sets is not more than 30 entity categories. However, the real-world entity class tree is much larger than 30, and the conventional idea is how many entity classes label how much data, which is not practical. It is often the case that when a new entity class appears, there are often only 10 to 100 samples of the new class, and it is not practical to retrain the model with these samples, since the model must be affected by class imbalance and overfitting.
Therefore, there is a need for a method that can accurately identify the corresponding entity class in the data when a new entity class appears.
Disclosure of Invention
In view of the above, it is necessary to provide a meta-learning based entity class identification method, apparatus, device and storage medium capable of accurately identifying entity classes in view of the above technical problems.
A meta-learning based entity class identification method, the method comprising:
acquiring a newly added entity type, and inquiring a reference sample corresponding to the newly added entity type;
acquiring data to be identified;
and inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model so as to recognize a new entity class corresponding to the reference sample in each data to be recognized, wherein the entity class recognition model is obtained by training based on a meta-learning mode.
In one embodiment, the inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model to recognize a new entity class corresponding to the reference sample in each data to be recognized includes:
serializing words in the reference sample and the data to be recognized, and performing high-order feature representation on the serialized words;
carrying out average pooling operation on the words after the high-order feature representation to obtain vector representations of the reference sample and the data to be identified;
processing the vectorization representation of the data to be identified through the vectorization representation of the reference sample to obtain high-level features of the data to be identified;
and processing the high-level features to obtain the newly added entity category corresponding to the reference sample in the data to be identified.
In one embodiment, the training mode of the entity class identification model includes:
obtaining sample data, and constructing a multi-element training sample according to the sample data;
and training according to the meta-training sample to obtain an entity class identification model.
In one embodiment, the obtaining sample data and constructing a multi-element training sample according to the sample data includes:
obtaining sample data, grouping the sample data according to entity types, and randomly extracting at least one group from the groups;
determining a first quantity of sample data in the extracted at least one group as a support sample, and determining a second quantity of sample data as a query sample;
obtaining a component training sample according to the supporting sample and the query sample;
repeating the randomly extracting at least one packet from the packets to obtain a multi-element training sample.
In one embodiment, the obtaining sample data and grouping the sample data according to entity categories includes:
obtaining sample data grouped according to an initial entity type, and grouping the sample data in the initial entity type according to a target entity type;
carrying out standardization processing on sample data grouped according to the target entity type;
and merging the standardized target entity categories corresponding to the initial entity categories to obtain groups corresponding to the target entity categories.
In one embodiment, the training according to the meta-training sample to obtain the entity class recognition model includes:
serializing words in the support sample and the query sample, performing high-order feature representation on the serialized words, and performing average pooling operation on the words after the high-order feature representation to obtain vector representations of the support sample and the query sample;
processing the vectorization representation of the query sample through the vectorization representation of the support sample according to an entity category identification model to obtain the high-level features of the query sample;
processing the high-level characteristics of the query sample to obtain a newly added entity category corresponding to the support sample in the query sample;
inputting the newly added entity type corresponding to the support sample in the obtained query sample and the real entity type of the query sample into a random field layer to obtain a loss function through calculation;
and training the entity class recognition model through the loss function.
In one embodiment, the serializing the words in the support sample and the query sample, and performing high-order feature representation on the serialized words includes:
processing the vectorized representation of the query sample by the vectorized representation of the support sample according to the following formula to obtain the high-level features of the query sample:
Figure RE-GDA0002910759790000031
Figure RE-GDA0002910759790000032
wherein the content of the first and second substances,
Figure RE-GDA0002910759790000033
is qjHigh-level features of the query sample obtained after modeling of the support sample; the atten function is used for calculating the contribution degree of each support sample to the identification of the named entity of the query sample;
Figure RE-GDA0002910759790000034
the method is characterized in that two vectors are spliced into a new vector, T is a real number and is used for controlling the sharpness degree of distribution obtained by an atten function; k represents the serial number of the supporting samples, and the value of k is related to the number of the supporting samples.
An entity class identification apparatus based on meta-learning, the apparatus comprising:
the new entity type acquisition module is used for acquiring a new entity type and inquiring a reference sample corresponding to the new entity type;
the data to be identified acquisition module is used for acquiring data to be identified;
and the entity identification module is used for inputting the reference sample and the data to be identified into a pre-generated entity class identification model so as to identify a newly added entity class corresponding to the reference sample in each data to be identified, wherein the entity class identification model is obtained by training based on a meta-learning mode.
A computer device comprising a memory storing a computer program and a processor implementing the steps of the method in any of the above embodiments when the processor executes the computer program.
A computer storage medium having stored thereon a computer program which, when executed by a processor, carries out the steps of the method in any of the above embodiments.
According to the entity category identification method, the entity category identification device, the entity category identification equipment and the storage medium based on meta-learning, the reference sample is determined according to the newly added entity category, the reference sample and the data to be identified are input into the entity category identification model which is generated in advance, so that the newly added entity category corresponding to the reference sample in each data to be identified is identified, manual interference is not needed, special knowledge in the field of artificial intelligence is not needed, the labor cost is greatly reduced, when the newly added entity category exists, the model does not need to be retrained, and only a few reference samples are needed to identify the data to be identified so as to determine whether the entity category exists.
Drawings
FIG. 1 is a flow diagram illustrating a method for entity class identification based on meta-learning in one embodiment;
FIG. 2 is a flow chart illustrating a meta-learning based entity class identification method according to another embodiment;
FIG. 3 is a block diagram of an entity class identification apparatus based on meta-learning in one embodiment;
FIG. 4 is a diagram illustrating an internal structure of a computer device according to an embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, the present application is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the present application and are not intended to limit the present application.
In an embodiment, as shown in fig. 1, an entity category identification method based on meta learning is provided, and this embodiment is illustrated by applying this method to a terminal, it is to be understood that this method may also be applied to a server, and may also be applied to a system including a terminal and a server, and is implemented by interaction between the terminal and the server. In this embodiment, the method includes the steps of:
s102: and acquiring the new entity type, and inquiring a reference sample corresponding to the new entity type.
Specifically, the new entity category may be a name of a new entity, and the new entity category may be at least one. The reference samples are samples belonging to the newly added entity category, wherein the number of the reference samples may be 10 or more, but the number is not too large. The server may establish a correspondence between the newly added entity class and the corresponding reference sample, for example, in a grouping manner. In addition, it should be noted that the same reference sample may belong to a plurality of additional entity categories, that is, one reference sample may be labeled with a plurality of entity categories.
S104: and acquiring data to be identified.
Specifically, the data to be identified is data that needs to be subjected to entity classification processing, and may be newly added data or past data.
S106: and inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model so as to recognize a new entity class corresponding to the reference sample in each data to be recognized, wherein the entity class recognition model is obtained by training based on a meta-learning mode.
Specifically, the entity class identification model is obtained by training based on a meta-learning mode, wherein a plurality of meta-training tasks are constructed according to sample data, and then the entity class identification model is obtained by training through the constructed meta-training tasks. The meta-training task is an entity class identification model which can identify the entity class of the data to be identified by training a few newly added entity class samples obtained by giving a small number of support samples and a large number of query samples.
The server inputs the reference sample and the data to be identified into a pre-generated entity type identification model, so that the entity type identification model identifies a newly added entity type corresponding to the reference sample in each data to be identified.
The process of identifying the newly added entity type corresponding to the reference sample in each piece of data to be identified by the entity type identification model may include a process of processing the reference sample and the piece of data to be identified, a step of calculating a high-level feature representation of the piece of data to be identified by using the processed reference sample, and a step of processing according to the high-level feature representation to determine the newly added entity type corresponding to the reference sample in each piece of data to be identified.
The process of processing the reference sample and the data to be identified may include: and finally, processing the words after the high-order representation through average pooling operation to obtain vector representations corresponding to the reference sample and the data to be recognized.
The step of the server calculating the high-level feature representation of the data to be identified by means of the processed reference samples may be performed according to the following formula:
Figure RE-GDA0002910759790000061
Figure RE-GDA0002910759790000062
wherein the content of the first and second substances,
Figure RE-GDA0002910759790000063
is qjThe high-level features of the data to be identified, which are obtained after modeling by the reference sample, model the relationship between the reference sample and the data to be identified to some extent. The atten function is used for calculating the contribution degree of each reference sample to the named entity identification in the data to be identified.
Figure RE-GDA0002910759790000064
Representing the concatenation of two vectors into a new longer vector, T is a real number that controls how sharp the atten-derived distribution is. k represents the index of the reference sample, which is related to the number of samples of the reference sample.
Finally, the step of processing according to the high-level feature representation to determine the newly added entity category of the corresponding reference sample in each data to be identified includes: inputting high-level features into a predetermined full-connected layer, mapping the feature vector dimension of each word to a preset dimension, for example, 3 dimensions, which respectively represent that label of the word is O, B, I, i.e. does not belong to the category, belongs to the category and is located at the beginning of the sentence, belongs to the category and is located in the middle of the sentence.
It should be emphasized that, in order to further ensure the privacy and security of the entity types of the newly added entity and the data to be identified, the entity types of the newly added entity and the data to be identified may also be stored in a node of a block chain.
According to the entity category identification method based on meta-learning, the reference sample is determined according to the newly added entity category, the reference sample and the data to be identified are input into the entity category identification model generated in advance, the newly added entity category corresponding to the reference sample in each data to be identified is identified, manual interference is not needed, knowledge in the special artificial intelligence field is not needed, labor cost is greatly reduced, when the newly added entity category exists, the model does not need to be retrained, and only a few reference samples are needed to identify the data to be identified so as to determine whether the entity category exists.
In one embodiment, inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model to recognize a new entity class corresponding to the reference sample in each data to be recognized includes: serializing words in the reference sample and the data to be recognized, and performing high-order feature representation on the serialized words; carrying out average pooling operation on the words after the high-order feature representation to obtain vector representations of the reference sample and the data to be identified; processing the vectorization representation of the data to be identified by referring to the vectorization representation of the sample to obtain the high-level features of the data to be identified; and processing the high-level features to obtain the newly added entity category of the corresponding reference sample in the data to be identified.
Specifically, assume that the word sequences of the reference samples are respectively
Figure RE-GDA0002910759790000071
Then the input of the reference sample is [ CLS ]],
Figure RE-GDA0002910759790000072
[SEP]The word sequence of the data to be recognized is
Figure RE-GDA0002910759790000073
The input of the data to be recognized is [ CLS ]],
Figure RE-GDA0002910759790000074
[SEP]。
After the reference sample and the data to be recognized are input into the BERT, the high-order feature representation of each word of the reference sample and the data to be recognized can be obtained by the following equation:
Figure RE-GDA0002910759790000075
Figure RE-GDA0002910759790000076
wherein
Figure RE-GDA0002910759790000077
And
Figure RE-GDA0002910759790000078
i and j words, s, of reference sample and data to be recognized, respectivelyiAnd q isjRespectively, the high-order feature representations of the two words after BERT.
After obtaining the high-order feature representations of the words, an average pooling operation is used to obtain a uniform vector representation representing the entire reference sample and the data to be recognized:
srep=MEAN_POOLINGi(si)
qrep=MEAN_POOLINGj(qj)
s thus obtainedrepThen generationTable characterization of the entire reference sample, qrepIt represents a characteristic representation of the entire data to be identified.
After obtaining the feature representations of the entire reference sample and the data to be identified, a higher-order feature representation of the data to be identified may be obtained from the reference sample:
Figure RE-GDA0002910759790000079
Figure RE-GDA0002910759790000081
wherein
Figure RE-GDA0002910759790000082
Is qjThe high-level features of the data to be identified, which are obtained after modeling by the reference sample, model the relationship between the reference sample and the data to be identified to some extent. The atten function is used for calculating the contribution degree of each reference sample to the named entity identification in the data to be identified.
Figure RE-GDA0002910759790000083
Representing the concatenation of two vectors into a new longer vector, T is a real number that controls the sharpness of the distribution obtained by the atten function. k represents the index of the reference sample, which is related to the number of samples of the reference sample.
The server obtains the final feature representation of each word in the query sample, and then maps the feature vector dimension of each word to 3 dimensions through a full connection layer, wherein the three dimensions respectively represent that the label of the word is O, B and I, namely the label does not belong to the category, belongs to the category and is positioned at the beginning of a sentence, belongs to the category and is positioned in the middle of the sentence, namely the new entity category corresponding to the reference sample in the data to be recognized.
In one embodiment, the training mode of the entity class recognition model includes: acquiring sample data, and constructing a multi-element training sample according to the sample data; and training according to the meta-training sample to obtain an entity class identification model.
Specifically, the sample data may be a sample that is set in advance and has been classified. The meta training samples are obtained by processing sample data, wherein each meta training sample may include a plurality of support samples and a plurality of query samples, wherein the support samples may include a plurality of groups of sample data, i.e., sample data belonging to different categories, and the corresponding query sample is also the query sample in the corresponding group. The number of groups of the meta-training samples can be set according to needs, for example, ten thousand, then the meta-training samples are trained to obtain a target classification model, for example, the meta-training samples are trained sequentially until the accuracy of the entity class identification model reaches an expectation, wherein the calculation of the accuracy of the entity class identification model can be processed according to the meta-training samples, for example, the support samples and the query samples in the meta-training samples are input into the entity class identification model to determine the entity class corresponding to the query sample, and if the accuracy is compared with the real entity class of the query sample and the expectation is reached, the model training is completed.
In one embodiment, obtaining sample data and constructing a multi-element training sample according to the sample data includes: obtaining sample data, grouping the sample data according to entity types, and randomly extracting at least one group from the groups; determining a first quantity of sample data in the extracted at least one group as a support sample, and determining a second quantity of sample data as a query sample; obtaining a component training sample according to the supporting sample and the query sample; at least one packet is repeatedly and randomly extracted from the packets to obtain a multi-element training sample.
Specifically, the server first obtains original sample data, then processes the original sample data to obtain a data set corresponding to each category, and then starts to construct a training set. In order to train the meta-learning model, a series of meta-training samples are first constructed, and the construction rule is as follows:
randomly extracting a number of groups, e.g. 3 classes, from the processed entity classes, e.g. 12 entity classesIs represented by1,l2,…,l3From1,l2,…,l3Of these 3 classes, a first number, e.g., 10 samples, are randomly drawn for each class as support samples, and a second number, e.g., 100 samples, are randomly drawn for each class as query samples, so that a total of 30 support samples, 3000 query samples, are obtained.
The server integrates the data thus constructed into a meta-training task (meta-training task) whose purpose is to train the model so that it can classify the query sample given the supporting sample. To train the model, the server may construct 10000 such meta-training tasks.
In one embodiment, obtaining sample data, and grouping the sample data according to entity categories includes: obtaining sample data grouped according to the initial entity category, and grouping the sample data in the initial entity category according to the target entity category; carrying out standardization processing on sample data grouped according to the target entity type; and merging the standardized target entity categories corresponding to the initial entity categories to obtain groups corresponding to the target entity categories.
Specifically, the initial entity category is open source MSRA, civil newspaper, microblog, clener, BOSON data sets collected from the internet, and since the labeling formats of the data sets are not uniform, the data are preprocessed to be unified into data in a BIO labeling format.
Wherein the target entity is labeled in the initial category entity, such as PER name, LOC location, ORG organization, TIM time, COM company name, ADD specific address, GAME GAME name, GOV government department, SCENCE attraction, BOOK BOOK, MOVIE MOVIE, and PRODUCT PRODUCT. The server counts entity types marked by each data set in the MSRA, the civil daily report, the microblog, the CLUENER and the BOSON data sets, wherein the MSRA marks three entity types including PER, LOC and ORG, and if L (MSRA) is the entity type set marked by the MSRA, L (MSRA) ═ PER, LOC and ORG. Similarly, the server may obtain L (people's daily report) { PER, LOC, ORG, TIM }, L { microblog } { PER, ORG, LOC }, L (timer) { PER, LOC, ORG, COM, ADD, GAME, GOV, SCENCE, BOOK, MOVIE }, L (guest) { PER, LOC, ORG, COM, TIM, PRODUCT }
According to the data set labeling entity class set obtained in the last step, each data set is divided into new entity classes according to a single entity class, for example, for an MSRA data set, l (MSRA) ═ { PER, LOC, ORG }, the PER class is considered first, all PER positive samples in the MSRA data set are retained, all other positive samples such as LOC and ORG are labeled as negative samples, the originally existing negative samples in the MSRA remain unchanged, only the PER class positive samples are contained in the newly obtained data set, all other classes positive samples become negative samples, and finally sentences all of which are negative samples are removed, the data set is recorded as MSRA-PER, and similarly, the server can obtain MSRA-ORG, MSRA-LOC data sets. The server may also go to CLUENER-PER, CLUENER-ADD. Specifically, i call AB, live on CD, and work on EF, for example. Wherein AB is PER entity, CD is LOC entity, EF is ORG entity, these all count positive samples, and I call, live, these all count negative samples in office.
Then, the server can obtain five data sets related to the PERs, namely MSRA-PER, people's daily report-PER, CLUENER-PER, microblog-PER and BOSON-PER, and through the analysis, the five data sets only contain the entity of the PER category, and the entities of other categories are negative samples, so that the five data sets related to the PER are mixed to form a new data set which is recorded as a ZH-PER data set. Similarly, the server can go to ZH-LOC, ZH-ORG, ZH-TIM, ZH-ADD, ZH-COM, ZH-BOOK, etc. for a total of 12 data sets.
In one embodiment, referring to fig. 2, fig. 2 is a flowchart of a training process of the entity class identification model in an embodiment, where the training according to the meta-training sample to obtain the entity class identification model includes: serializing words in the support sample and the query sample, performing high-order feature representation on the serialized words, and performing average pooling operation on the words after the high-order feature representation to obtain vector representations of the support sample and the query sample; processing the vectorization representation of the query sample through the vectorization representation of the support sample according to the entity type identification model to obtain the high-level characteristics of the query sample; processing the high-level characteristics of the query sample to obtain a newly added entity type of a corresponding support sample in the query sample; inputting the newly added entity category of the corresponding support sample in the obtained query sample and the real entity category of the query sample into a random field layer to calculate and obtain a loss function; and training the entity class recognition model through a loss function.
Specifically, after the meta-training task is built, the server begins building the model. The text adopts a Chinese pre-training language model BERT to code the characteristic representation of a sentence, and the main structure of the model is as follows:
let the word sequences of the support (support) samples be respectively
Figure RE-GDA0002910759790000111
The input to support the sample is [ CLS ]],
Figure RE-GDA0002910759790000112
[SEP]The word sequence of the query sample is
Figure RE-GDA0002910759790000113
The input of the query sample is [ CLS ]],
Figure RE-GDA0002910759790000114
[SEP]。
After entering the support and query samples into BERT, the server obtains a high-order feature representation for each word of these samples by:
Figure RE-GDA0002910759790000115
Figure RE-GDA0002910759790000116
wherein
Figure RE-GDA0002910759790000117
And
Figure RE-GDA0002910759790000118
the ith and jth words, s, of the support and query samples, respectivelyiAnd q isjRespectively, the high-order feature representations of the two words after BERT.
After obtaining the high-order feature representations of the words, the server uses an averaging pooling operation to obtain a uniform vector representation representing the entire sample:
srep=MEAN_POOLINGi(si)
qrep=MEAN_POOLINGj(qj)
s thus obtainedrepRepresenting the feature representation of the entire support sample, qrepA characterization representation of the entire query sample is represented.
After obtaining the feature representation of the entire sample, the server obtains a higher-order feature representation of the query sample from the support sample:
Figure RE-GDA0002910759790000119
wherein
Figure RE-GDA00029107597900001110
Is qjThe higher-level feature representation obtained after modeling the support sample models to some extent the relationship between the support sample and the query sample. The atten function is used for calculating the contribution degree of each support sample to the identification of the named entity of the query sample.
Figure RE-GDA00029107597900001111
Representing the concatenation of two vectors into a new longer vector, T is a real number that controls how sharp the atten-derived distribution is. k represents the serial number of the support samples, since 10 support samples were chosen for each category, since k takes 10 at maximum.
Thus, the server obtains the final feature representation of each word in the query sample, and then maps the feature vector dimension of each word to 3 dimensions through a full connection layer, where the three dimensions represent the words, label, to be O, B, I, i.e. not belonging to the category, belonging to the category and located at the beginning of the sentence, and belonging to the category and located in the middle of the sentence. After each word is mapped to 3 dimensions through the fully connected layer, a conditional random field CRF layer is then used to calculate the final penalty. The model is trained by a loss function.
In one embodiment, serializing the words in the support sample and the query sample, and performing high-order feature representation on the serialized words includes: processing the vectorized representation of the query sample by supporting the vectorized representation of the sample according to the following formula to obtain the high-level features of the query sample:
Figure RE-GDA0002910759790000121
Figure RE-GDA0002910759790000122
wherein the content of the first and second substances,
Figure RE-GDA0002910759790000123
is qjHigh-level features of a query sample obtained after modeling of the support sample; the atten function is used for calculating the contribution degree of each support sample to the identification of the named entity of the query sample;
Figure RE-GDA0002910759790000124
represents twoSplicing the vectors into a new vector, wherein T is a real number and is used for controlling the sharpness degree of the distribution obtained by the atten function; k represents the serial number of the supporting samples, and the value of k is related to the number of the supporting samples.
It should be understood that although the various steps in the flowcharts of fig. 1 and 2 are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and described, and may be performed in other orders, unless explicitly stated otherwise. Moreover, at least some of the steps in fig. 1 and 2 may include multiple sub-steps or multiple stages that are not necessarily performed at the same time, but may be performed at different times, and the order of performing the sub-steps or stages is not necessarily sequential, but may be performed in turn or alternately with other steps or at least some of the sub-steps or stages of other steps.
In one embodiment, as shown in fig. 3, there is provided a meta-learning based entity class identification apparatus, including: an entity category acquisition module 100, a data to be identified acquisition module 200 and an entity identification module 300 are added, wherein:
a new entity type obtaining module 100, configured to obtain a new entity type, and query a reference sample corresponding to the new entity type;
a to-be-identified data acquisition module 200, configured to acquire to-be-identified data;
the entity identification module 300 is configured to input the reference sample and the data to be identified into a pre-generated entity class identification model, so as to identify a new entity class corresponding to the reference sample in each data to be identified, where the entity class identification model is obtained by training based on meta-learning.
In one embodiment, the entity identification module 300 may include:
the conversion unit is used for serializing the words in the reference sample and the data to be identified and performing high-order feature representation on the serialized words;
the first vector quantization unit is used for carrying out average pooling operation on the words after the high-order characteristic representation to obtain vector representations of a reference sample and data to be identified;
the first high-level feature representation unit is used for processing the vectorization representation of the data to be identified by referring to the vectorization representation of the sample to obtain the high-level features of the data to be identified;
and the identification unit is used for processing the high-level characteristics to obtain the newly added entity type of the corresponding reference sample in the data to be identified.
In one embodiment, the entity class identification device based on meta-learning includes:
the sample acquisition module is used for acquiring sample data and constructing a multi-element training sample according to the sample data;
and the training is fast, and the entity class recognition model is obtained by training according to the meta-training sample.
In one embodiment, the sample acquiring module may include:
the grouping unit is used for acquiring sample data, grouping the sample data according to the entity type and randomly extracting at least one group from the groups;
the extraction unit is used for determining that the first quantity of sample data in the extracted at least one group is a support sample and the second quantity of sample data is a query sample;
the combination unit is used for obtaining a group of element training samples according to the support samples and the query samples;
and the circulating unit is used for repeatedly and randomly extracting at least one packet from the packets to obtain the multi-element training sample.
In one embodiment, the grouping unit may include:
the grouping subunit is used for acquiring sample data grouped according to the initial entity category and grouping the sample data in the initial entity category according to the target entity category;
the standardization subunit is used for carrying out standardization processing on the sample data grouped according to the target entity type;
and the merging subunit is used for merging the standardized target entity categories corresponding to the initial entity categories to obtain the groups corresponding to the target entity categories.
In one embodiment, the training module may include:
the second vector quantization unit is used for serializing the words in the support sample and the query sample, performing high-order feature representation on the serialized words, and performing average pooling operation on the words after the high-order feature representation to obtain vector representations of the support sample and the query sample;
the second high-level feature representation unit is used for processing the vectorization representation of the query sample through the vectorization representation of the support sample according to the entity type identification model to obtain the high-level features of the query sample;
the category identification unit is used for processing the high-level characteristics of the query sample to obtain the category of a newly added entity corresponding to the support sample in the query sample;
the loss function generating unit is used for inputting the newly added entity type of the corresponding support sample in the obtained query sample and the real entity type of the query sample into the random field layer to calculate and obtain a loss function;
and the training unit is used for training the entity class recognition model through a loss function.
In one embodiment, the second quantization unit is further configured to process the vectorized representation of the query sample by supporting the vectorized representation of the sample according to the following formula to obtain the high-level features of the query sample:
Figure RE-GDA0002910759790000141
Figure RE-GDA0002910759790000142
wherein the content of the first and second substances,
Figure RE-GDA0002910759790000143
is qjHigh-level features of a query sample obtained after modeling of the support sample; the atten function is used for calculating the contribution degree of each support sample to the identification of the named entity of the query sample;
Figure RE-GDA0002910759790000144
the method is characterized in that two vectors are spliced into a new vector, T is a real number and is used for controlling the sharpness degree of distribution obtained by an atten function; k represents the serial number of the supporting samples, and the value of k is related to the number of the supporting samples.
For the specific definition of the entity class identification device based on meta-learning, reference may be made to the above definition of the entity class identification method based on meta-learning, which is not described herein again. The modules in the entity category identification device based on meta-learning can be wholly or partially realized by software, hardware and a combination thereof. The modules can be embedded in a hardware form or independent from a processor in the computer device, and can also be stored in a memory in the computer device in a software form, so that the processor can call and execute operations corresponding to the modules.
In one embodiment, a computer device is provided, which may be a server, the internal structure of which may be as shown in fig. 4. The computer device includes a processor, a memory, a network interface, and a database connected by a system bus. Wherein the processor of the computer device is configured to provide computing and control capabilities. The memory of the computer device comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system, a computer program, and a database. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The network interface of the computer device is used for communicating with an external terminal through a network connection. The computer program is executed by a processor to implement a meta-learning based entity class identification method.
Those skilled in the art will appreciate that the architecture shown in fig. 4 is merely a block diagram of some of the structures associated with the disclosed aspects and is not intended to limit the computing devices to which the disclosed aspects apply, as particular computing devices may include more or less components than those shown, or may combine certain components, or have a different arrangement of components.
In one embodiment, there is provided a computer device comprising a memory storing a computer program and a processor implementing the following steps when the processor executes the computer program: acquiring a new entity type, and inquiring a reference sample corresponding to the new entity type; acquiring data to be identified; and inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model so as to recognize a new entity class corresponding to the reference sample in each data to be recognized, wherein the entity class recognition model is obtained by training based on a meta-learning mode.
In one embodiment, the inputting of the reference sample and the data to be recognized into the entity class recognition model generated in advance to recognize the new entity class corresponding to the reference sample in each data to be recognized, which is realized when the processor executes the computer program, includes: serializing words in the reference sample and the data to be recognized, and performing high-order feature representation on the serialized words; carrying out average pooling operation on the words after the high-order feature representation to obtain vector representations of the reference sample and the data to be identified; processing the vectorization representation of the data to be identified by referring to the vectorization representation of the sample to obtain the high-level features of the data to be identified; and processing the high-level features to obtain the newly added entity category of the corresponding reference sample in the data to be identified.
In one embodiment, the manner in which the entity class identification model is trained when the computer program is executed by the processor comprises: acquiring sample data, and constructing a multi-element training sample according to the sample data; and training according to the meta-training sample to obtain an entity class identification model.
In one embodiment, obtaining sample data and constructing a multi-component training sample from the sample data, as implemented by a processor executing a computer program, comprises: obtaining sample data, grouping the sample data according to entity types, and randomly extracting at least one group from the groups; determining a first quantity of sample data in the extracted at least one group as a support sample, and determining a second quantity of sample data as a query sample; obtaining a component training sample according to the supporting sample and the query sample; at least one packet is repeatedly and randomly extracted from the packets to obtain a multi-element training sample.
In one embodiment, obtaining sample data implemented by a processor executing a computer program, grouping the sample data by entity class, comprises: obtaining sample data grouped according to the initial entity category, and grouping the sample data in the initial entity category according to the target entity category; carrying out standardization processing on sample data grouped according to the target entity type; and merging the standardized target entity categories corresponding to the initial entity categories to obtain groups corresponding to the target entity categories.
In one embodiment, the training according to the meta-training samples to obtain the entity class recognition model when the processor executes the computer program includes: serializing words in the support sample and the query sample, performing high-order feature representation on the serialized words, and performing average pooling operation on the words after the high-order feature representation to obtain vector representations of the support sample and the query sample; processing the vectorization representation of the query sample through the vectorization representation of the support sample according to the entity type identification model to obtain the high-level characteristics of the query sample; processing the high-level characteristics of the query sample to obtain a newly added entity type of a corresponding support sample in the query sample; inputting the newly added entity category of the corresponding support sample in the obtained query sample and the real entity category of the query sample into a random field layer to calculate and obtain a loss function; and training the entity class recognition model through a loss function.
In one embodiment, the serialization of words in the support sample and the query sample and the high-order feature representation of the serialized words, which are implemented by the processor when the computer program is executed, comprises: processing the vectorized representation of the query sample by supporting the vectorized representation of the sample according to the following formula to obtain the high-level features of the query sample:
Figure RE-GDA0002910759790000161
Figure RE-GDA0002910759790000162
wherein the content of the first and second substances,
Figure RE-GDA0002910759790000171
is qjHigh-level features of a query sample obtained after modeling of the support sample; the atten function is used for calculating the contribution degree of each support sample to the identification of the named entity of the query sample;
Figure RE-GDA0002910759790000172
the method is characterized in that two vectors are spliced into a new vector, T is a real number and is used for controlling the sharpness degree of distribution obtained by an atten function; k represents the serial number of the supporting samples, and the value of k is related to the number of the supporting samples.
In one embodiment, a computer storage medium is provided, having a computer program stored thereon, the computer program, when executed by a processor, implementing the steps of: acquiring a new entity type, and inquiring a reference sample corresponding to the new entity type; acquiring data to be identified; and inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model so as to recognize a new entity class corresponding to the reference sample in each data to be recognized, wherein the entity class recognition model is obtained by training based on a meta-learning mode.
In one embodiment, the inputting of the reference sample and the data to be recognized into the entity class recognition model generated in advance to recognize the new entity class corresponding to the reference sample in each data to be recognized, which is realized when the computer program is executed by the processor, includes: serializing words in the reference sample and the data to be recognized, and performing high-order feature representation on the serialized words; carrying out average pooling operation on the words after the high-order feature representation to obtain vector representations of the reference sample and the data to be identified; processing the vectorization representation of the data to be identified by referring to the vectorization representation of the sample to obtain the high-level features of the data to be identified; and processing the high-level features to obtain the newly added entity category of the corresponding reference sample in the data to be identified.
In one embodiment, the manner in which the entity class identification model is trained when the computer program is executed by the processor includes: acquiring sample data, and constructing a multi-element training sample according to the sample data; and training according to the meta-training sample to obtain an entity class identification model.
In one embodiment, obtaining sample data and constructing a multi-component training sample from the sample data, when the computer program is executed by a processor, comprises: obtaining sample data, grouping the sample data according to entity types, and randomly extracting at least one group from the groups; determining a first quantity of sample data in the extracted at least one group as a support sample, and determining a second quantity of sample data as a query sample; obtaining a component training sample according to the supporting sample and the query sample; at least one packet is repeatedly and randomly extracted from the packets to obtain a multi-element training sample.
In one embodiment, obtaining sample data, the grouping of sample data by entity class performed when the computer program is executed by the processor, comprises: obtaining sample data grouped according to the initial entity category, and grouping the sample data in the initial entity category according to the target entity category; carrying out standardization processing on sample data grouped according to the target entity type; and merging the standardized target entity categories corresponding to the initial entity categories to obtain groups corresponding to the target entity categories.
In one embodiment, training from meta-training samples to obtain an entity class recognition model, when implemented by a processor, comprises: serializing words in the support sample and the query sample, performing high-order feature representation on the serialized words, and performing average pooling operation on the words after the high-order feature representation to obtain vector representations of the support sample and the query sample; processing the vectorization representation of the query sample through the vectorization representation of the support sample according to the entity type identification model to obtain the high-level characteristics of the query sample; processing the high-level characteristics of the query sample to obtain a newly added entity type of a corresponding support sample in the query sample; inputting the newly added entity category of the corresponding support sample in the obtained query sample and the real entity category of the query sample into a random field layer to calculate and obtain a loss function; and training the entity class recognition model through a loss function.
In one embodiment, the serialization of words in the support and query samples and the high-order feature representation of the serialized words, implemented when the computer program is executed by a processor, includes: processing the vectorized representation of the query sample by supporting the vectorized representation of the sample according to the following formula to obtain the high-level features of the query sample:
Figure RE-GDA0002910759790000181
Figure RE-GDA0002910759790000182
wherein the content of the first and second substances,
Figure RE-GDA0002910759790000183
is qjHigh-level features of a query sample obtained after modeling of the support sample; the atten function is used for calculating the contribution degree of each support sample to the identification of the named entity of the query sample;
Figure RE-GDA0002910759790000184
the method is characterized in that two vectors are spliced into a new vector, T is a real number and is used for controlling the sharpness degree of distribution obtained by an atten function; k represents the serial number of the supporting sample, and the value of k and the supporting sampleIs correlated with the number of samples.
The block chain is a novel application mode of computer technologies such as distributed data storage, point-to-point transmission, a consensus mechanism, an encryption algorithm and the like. A block chain (Blockchain), which is essentially a decentralized database, is a series of data blocks associated by using a cryptographic method, and each data block contains information of a batch of network transactions, so as to verify the validity (anti-counterfeiting) of the information and generate a next block. The blockchain may include a blockchain underlying platform, a platform product service layer, an application service layer, and the like.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware instructions of a computer program, which can be stored in a non-volatile computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. Any reference to memory, storage, database, or other medium used in the embodiments provided herein may include non-volatile and/or volatile memory, among others. Non-volatile memory can include read-only memory (ROM), Programmable ROM (PROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), Rambus Direct RAM (RDRAM), direct bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
The technical features of the above embodiments can be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the above embodiments are not described, but should be considered as the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
The above-mentioned embodiments only express several embodiments of the present application, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the concept of the present application, which falls within the scope of protection of the present application. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (10)

1. A meta-learning based entity class identification method, the method comprising:
acquiring a newly added entity type, and inquiring a reference sample corresponding to the newly added entity type;
acquiring data to be identified;
and inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model so as to recognize a new entity class corresponding to the reference sample in each data to be recognized, wherein the entity class recognition model is obtained by training based on a meta-learning mode.
2. The method of claim 1, wherein the inputting the reference sample and the data to be recognized into a pre-generated entity class recognition model to recognize a new entity class corresponding to the reference sample in each data to be recognized comprises:
serializing words in the reference sample and the data to be recognized, and performing high-order feature representation on the serialized words;
carrying out average pooling operation on the words after the high-order feature representation to obtain vector representations of the reference sample and the data to be identified;
processing the vectorization representation of the data to be identified through the vectorization representation of the reference sample to obtain high-level features of the data to be identified;
and processing the high-level features to obtain the newly added entity category corresponding to the reference sample in the data to be identified.
3. The method according to claim 1 or 2, wherein the training of the entity class recognition model comprises:
obtaining sample data, and constructing a multi-element training sample according to the sample data;
and training according to the meta-training sample to obtain an entity class identification model.
4. The method of claim 3, wherein obtaining sample data and constructing multi-element training samples from the sample data comprises:
obtaining sample data, grouping the sample data according to entity types, and randomly extracting at least one group from the groups;
determining a first quantity of sample data in the extracted at least one group as a support sample, and determining a second quantity of sample data as a query sample;
obtaining a component training sample according to the supporting sample and the query sample;
repeating the randomly extracting at least one packet from the packets to obtain a multi-element training sample.
5. The method of claim 4, wherein the obtaining sample data, the grouping the sample data by entity class, comprises:
obtaining sample data grouped according to an initial entity type, and grouping the sample data in the initial entity type according to a target entity type;
carrying out standardization processing on sample data grouped according to the target entity type;
and merging the standardized target entity categories corresponding to the initial entity categories to obtain groups corresponding to the target entity categories.
6. The method of claim 5, wherein the training according to the meta-training samples to obtain an entity class recognition model comprises:
serializing words in the support sample and the query sample, performing high-order feature representation on the serialized words, and performing average pooling operation on the words after the high-order feature representation to obtain vector representations of the support sample and the query sample;
processing the vectorization representation of the query sample through the vectorization representation of the support sample according to an entity category identification model to obtain the high-level features of the query sample;
processing the high-level characteristics of the query sample to obtain a newly added entity category corresponding to the support sample in the query sample;
inputting the newly added entity type corresponding to the support sample in the obtained query sample and the real entity type of the query sample into a random field layer to obtain a loss function through calculation;
and training the entity class recognition model through the loss function.
7. The method of claim 6, wherein the serializing words in the support samples and the query samples and performing high-order feature representation on the serialized words comprises:
processing the vectorized representation of the query sample by the vectorized representation of the support sample according to the following formula to obtain the high-level features of the query sample:
Figure FDA0002836462110000021
Figure FDA0002836462110000022
wherein the content of the first and second substances,
Figure FDA0002836462110000023
is qjHigh-level features of the query sample obtained after modeling of the support sample; the atten function is used for calculating the contribution degree of each support sample to the identification of the named entity of the query sample;
Figure FDA0002836462110000031
the method is characterized in that two vectors are spliced into a new vector, T is a real number and is used for controlling the sharpness degree of distribution obtained by an atten function; k represents the serial number of the supporting samples, and the value of k is related to the number of the supporting samples.
8. An entity class identification device based on meta-learning, the device comprising:
the new entity type acquisition module is used for acquiring a new entity type and inquiring a reference sample corresponding to the new entity type;
the data to be identified acquisition module is used for acquiring data to be identified;
and the entity identification module is used for inputting the reference sample and the data to be identified into a pre-generated entity class identification model so as to identify a newly added entity class corresponding to the reference sample in each data to be identified, wherein the entity class identification model is obtained by training based on a meta-learning mode.
9. A computer device comprising a memory and a processor, the memory storing a computer program, wherein the processor implements the steps of the method of any one of claims 1 to 7 when executing the computer program.
10. A computer storage medium on which a computer program is stored, characterized in that the computer program, when being executed by a processor, carries out the steps of the method of any one of claims 1 to 7.
CN202011472865.XA 2020-12-15 2020-12-15 Entity category identification method, device, equipment and storage medium based on meta-learning Pending CN112528662A (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN202011472865.XA CN112528662A (en) 2020-12-15 2020-12-15 Entity category identification method, device, equipment and storage medium based on meta-learning
PCT/CN2021/109617 WO2022127124A1 (en) 2020-12-15 2021-07-30 Meta learning-based entity category recognition method and apparatus, device and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011472865.XA CN112528662A (en) 2020-12-15 2020-12-15 Entity category identification method, device, equipment and storage medium based on meta-learning

Publications (1)

Publication Number Publication Date
CN112528662A true CN112528662A (en) 2021-03-19

Family

ID=74999881

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011472865.XA Pending CN112528662A (en) 2020-12-15 2020-12-15 Entity category identification method, device, equipment and storage medium based on meta-learning

Country Status (2)

Country Link
CN (1) CN112528662A (en)
WO (1) WO2022127124A1 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022127124A1 (en) * 2020-12-15 2022-06-23 深圳壹账通智能科技有限公司 Meta learning-based entity category recognition method and apparatus, device and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101846824B1 (en) * 2017-12-11 2018-04-09 가천대학교 산학협력단 Automated Named-entity Recognizing Systems, Methods, and Computer-Readable Mediums
CN109783604A (en) * 2018-12-14 2019-05-21 平安科技(深圳)有限公司 Information extracting method, device and computer equipment based on a small amount of sample
CN110825875A (en) * 2019-11-01 2020-02-21 科大讯飞股份有限公司 Text entity type identification method and device, electronic equipment and storage medium
CN111797394A (en) * 2020-06-24 2020-10-20 广州大学 APT organization identification method, system and storage medium based on stacking integration
CN112001179A (en) * 2020-09-03 2020-11-27 平安科技(深圳)有限公司 Named entity recognition method and device, electronic equipment and readable storage medium
CN112052684A (en) * 2020-09-07 2020-12-08 南方电网数字电网研究院有限公司 Named entity identification method, device, equipment and storage medium for power metering

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109885825A (en) * 2019-01-07 2019-06-14 平安科技(深圳)有限公司 Name entity recognition method, device and computer equipment based on attention mechanism
CN111860580B (en) * 2020-06-09 2024-02-20 北京百度网讯科技有限公司 Identification model acquisition and category identification method, device and storage medium
CN111767400B (en) * 2020-06-30 2024-04-26 平安国际智慧城市科技股份有限公司 Training method and device for text classification model, computer equipment and storage medium
CN111859937A (en) * 2020-07-20 2020-10-30 上海汽车集团股份有限公司 Entity identification method and device
CN112528662A (en) * 2020-12-15 2021-03-19 深圳壹账通智能科技有限公司 Entity category identification method, device, equipment and storage medium based on meta-learning

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101846824B1 (en) * 2017-12-11 2018-04-09 가천대학교 산학협력단 Automated Named-entity Recognizing Systems, Methods, and Computer-Readable Mediums
CN109783604A (en) * 2018-12-14 2019-05-21 平安科技(深圳)有限公司 Information extracting method, device and computer equipment based on a small amount of sample
CN110825875A (en) * 2019-11-01 2020-02-21 科大讯飞股份有限公司 Text entity type identification method and device, electronic equipment and storage medium
CN111797394A (en) * 2020-06-24 2020-10-20 广州大学 APT organization identification method, system and storage medium based on stacking integration
CN112001179A (en) * 2020-09-03 2020-11-27 平安科技(深圳)有限公司 Named entity recognition method and device, electronic equipment and readable storage medium
CN112052684A (en) * 2020-09-07 2020-12-08 南方电网数字电网研究院有限公司 Named entity identification method, device, equipment and storage medium for power metering

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
YIMING ZHANG.ETC: "KADetector: Automatic Identification of Key Actors in Online Hack Forums Based on Structured Heterogeneous Information Network", IEEE, 31 December 2018 (2018-12-31) *
年福东;束建华;吕刚;: "基于自适应特征比较的少样本学习算法", 西安文理学院学报(自然科学版), no. 04, 15 October 2020 (2020-10-15) *
王浩畅;李钰;赵铁军;: "面向生物医学命名实体识别的多Agent元学习框架", 计算机学报, no. 07, 15 July 2010 (2010-07-15) *
王蕾;谢云;周俊生;顾彦慧;曲维光;: "基于神经网络的片段级中文命名实体识别", 中文信息学报, no. 03, 15 March 2018 (2018-03-15) *

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2022127124A1 (en) * 2020-12-15 2022-06-23 深圳壹账通智能科技有限公司 Meta learning-based entity category recognition method and apparatus, device and storage medium

Also Published As

Publication number Publication date
WO2022127124A1 (en) 2022-06-23

Similar Documents

Publication Publication Date Title
CN112613501A (en) Information auditing classification model construction method and information auditing method
CN109960725B (en) Text classification processing method and device based on emotion and computer equipment
CN109783785B (en) Method and device for generating experiment detection report and computer equipment
WO2022048363A1 (en) Website classification method and apparatus, computer device, and storage medium
CN109325118B (en) Unbalanced sample data preprocessing method and device and computer equipment
CN112052684A (en) Named entity identification method, device, equipment and storage medium for power metering
WO2022134586A1 (en) Meta-learning-based target classification method and apparatus, device and storage medium
CN112231416B (en) Knowledge graph body updating method and device, computer equipment and storage medium
CN112699923A (en) Document classification prediction method and device, computer equipment and storage medium
CN111914159A (en) Information recommendation method and terminal
CN114491084B (en) Self-encoder-based relation network information mining method, device and equipment
CN113656547A (en) Text matching method, device, equipment and storage medium
CN113254649B (en) Training method of sensitive content recognition model, text recognition method and related device
CN114095447B (en) Communication network encryption flow classification method based on knowledge distillation and self-distillation
CN113868419B (en) Text classification method, device, equipment and medium based on artificial intelligence
CN115730597A (en) Multi-level semantic intention recognition method and related equipment thereof
CN112528662A (en) Entity category identification method, device, equipment and storage medium based on meta-learning
CN110597977B (en) Data processing method, data processing device, computer equipment and storage medium
CN112199954A (en) Disease entity matching method and device based on voice semantics and computer equipment
CN116777646A (en) Artificial intelligence-based risk identification method, apparatus, device and storage medium
CN114266255B (en) Corpus classification method, apparatus, device and storage medium based on clustering model
CN112765183B (en) Multi-source data fusion method and device, storage medium and electronic equipment
CN112732423B (en) Process migration method, device, equipment and medium
CN109308565B (en) Crowd performance grade identification method and device, storage medium and computer equipment
CN112507170A (en) Data asset directory construction method based on intelligent decision and related equipment thereof

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
REG Reference to a national code

Ref country code: HK

Ref legal event code: DE

Ref document number: 40045440

Country of ref document: HK

SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination