CN109858041A - A kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries - Google Patents

A kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries Download PDF

Info

Publication number
CN109858041A
CN109858041A CN201910172675.7A CN201910172675A CN109858041A CN 109858041 A CN109858041 A CN 109858041A CN 201910172675 A CN201910172675 A CN 201910172675A CN 109858041 A CN109858041 A CN 109858041A
Authority
CN
China
Prior art keywords
model
training
custom dictionaries
lstm
semi
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201910172675.7A
Other languages
Chinese (zh)
Other versions
CN109858041B (en
Inventor
苏海波
高体伟
孙伟
王然
于帮付
黄伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baifendian Information Science & Technology Co Ltd
Original Assignee
Beijing Baifendian Information Science & Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baifendian Information Science & Technology Co Ltd filed Critical Beijing Baifendian Information Science & Technology Co Ltd
Priority to CN201910172675.7A priority Critical patent/CN109858041B/en
Publication of CN109858041A publication Critical patent/CN109858041A/en
Application granted granted Critical
Publication of CN109858041B publication Critical patent/CN109858041B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Character Discrimination (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses a kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries, include the following steps: S1, with unlabeled data pre-training Bi-LSTM language model;S2, use word vector model by each character vector at Embedding layers;S3, using two layers of two-way LSTM as sequence labelling model, the sequence labelling model is using labeled data training;S4, user's Custom Dictionaries are added;S5, maximum probability path in sequence is found out using Veterbi decoding.The present invention splices the output of the language model of pre-training and the output of the two-way LSTM of first layer work, and the input as the two-way LSTM of the second layer can reduce the use of mark corpus, while the mark corpus of frontier can be only replaced when switching field.In addition, when prediction the emission matrix into Veterbi decoding can be changed, to achieve the effect that Custom Dictionaries by the setting of Custom Dictionaries.

Description

A kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries
Technical field
The present invention relates to data processing fields, towards name entity recognition techniques application, and in particular to a kind of semi-supervised Practise the name entity recognition method for combining Custom Dictionaries.
Background technique
It names Entity recognition (Named Entity Recognition, hereinafter NER) to refer to and identifies tool from text There are the entity (usually noun), such as name, place name, mechanism name, proper noun etc. of particular category.NER is information retrieval, looks into The background task for the problems such as asking classification, automatic question answering, effect directly affect the effect of subsequent processing, therefore are at natural language Manage a underlying issue of research.
Semi-supervised learning (Semi-Supervised Learning, SSL) is pattern-recognition and machine learning area research Important Problems, be a kind of learning method that supervised learning is combined with unsupervised learning.Semi-supervised learning uses largely not Flag data, and flag data is used simultaneously, Lai Jinhang pattern-recognition work.The basic thought of semi-supervised learning is to utilize number Learner is established according to the model hypothesis in distribution, and label is carried out to non-label sample.Its formalized description is given one and comes from The sample set S=LU of certain unknown distribution, wherein L be label sample set L=(x1, y1), (x2, y2) ..., (x | L |, y | L |), U be a non-label sample set U=xc1, xc2 ..., xc | U | }, it is desirable to obtaining function f:XyY can be accurately to sample Example x predicts its label y.Wherein xi, xc1 are d dimensional vector, and ytIY is the label of sample xi, | L | and | U | it is respectively L's and U Size, that is, the sample number for being included, semi-supervised learning are exactly that optimal learner is found on sample set S.If S=L, Problem translates into traditional supervised learning;, whereas if S=U, then problem is to be converted into traditional unsupervised learning. Label sample and non-label sample are how comprehensively utilized, is semi-supervised learning problem to be solved.
Custom Dictionaries are the products based on user demand, and the user of different field and industry has different definition to entity And understanding, therefore the word having appears to be entity in certain user, and for other users, it may not be entity.Therefore user Custom Dictionaries be it is necessary, by dictionary can be improved name Entity recognition accuracy rate, so that it is more met user's Demand.
Summary of the invention
In view of the deficiencies of the prior art, the present invention is intended to provide a kind of name of semi-supervised learning combination Custom Dictionaries is real Body recognition methods.
In order to achieve the above technical purposes, the present invention adopts the following technical scheme:
A kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries, includes the following steps:
S1, with unlabeled data pre-training Bi-LSTM language model;
S2, use word vector model by each character vector at Embedding layers;
S3, using two layers of two-way LSTM as sequence labelling model, the sequence labelling model is instructed using labeled data Practice;
In the training process of sequence labelling model, by the output vector and step of the two-way LSTM of the first layer of sequence labelling model The output work for the Bi-LSTM language model that pre-training obtains in rapid S1 splices, then by the vector of splicing by a full connection Input after layer as the two-way LSTM of the second layer of sequence labelling model;
S4, user's Custom Dictionaries are added:
Emission matrix X can be obtained after two layers of two-way LSTM of sequence labelling model, by CRF layers, passes through maximum Likelihood probability obtains transfer matrix Y, and the probability of emission matrix, the hair after being adjusted then are adjusted according to user's Custom Dictionaries Penetrate matrix X;
S5, maximum probability path in sequence is found out using Veterbi decoding:
It will be input to obtained in step S4 according to user's Custom Dictionaries emission matrix X adjusted and transfer matrix Y CRF layers of Veterbi decoding obtains sequence labelling, i.e., correctly names Entity recognition result.
Further, in step S2, the word vector model is word2vec model.
Further, in step S2, the training of word vector model is specifically carried out using Skip-gram method.
Still further, carrying out the specific steps of word vector model training using Skip-gram method are as follows:
(1) the relevant balanced corpus of application field is collected first;
(2) corpus data collected for step (1) pre-processes, including filters out junk data, stops low-frequency word and nothing Meaning symbol, and it is organized into the format of training data, obtain training data;
(3) training data is given to Skip-gram model, training obtains word vector model.
The beneficial effects of the present invention are: it is based on pre-training language model (Pretrained Language Model), word Symbol insertion (char embeddings) technology, Custom Dictionaries technology, semi-supervised training (Semi-Supervised Learning), two-way LSTM (Long Short Term Memory, long memory models in short-term) network, CRF (Conditional Random Field, condition random field) model etc., the NER training of Lai Shixian semi-supervised learning.By above method and special Network structure splices the output of the language model of pre-training and the output of the two-way LSTM of first layer work, and two-way as the second layer The input of LSTM.Pass through the method, it is possible to reduce mark the use of corpus, while can only replace frontier when switching field Mark corpus.In addition, the transmitting into Veterbi decoding can be changed by the setting of Custom Dictionaries when prediction Matrix, to achieve the effect that Custom Dictionaries.
Detailed description of the invention
Fig. 1 is the method flow schematic diagram of the embodiment of the present invention;
Fig. 2 is the network diagram of Bi-LSTM language model in the embodiment of the present invention;
Fig. 3 is common word2vec training pattern CBOW schematic diagram in the embodiment of the present invention;
Fig. 4 is common word2vec training pattern skip-gram model schematic in the embodiment of the present invention;
Fig. 5 is the flow diagram of word vector model training in the embodiment of the present invention;
Fig. 6 is sequence labelling model schematic in the embodiment of the present invention.
Specific embodiment
Below with reference to attached drawing, the invention will be further described, it should be noted that the present embodiment is with this technology side Premised on case, the detailed implementation method and specific operation process are given, but protection scope of the present invention is not limited to this reality Apply example.
Simplicity of explanation is made to the technical term that the present embodiment is related to below:
Name Entity recognition: specific proper noun is identified from given text data, as name, place name, mechanism name, when Between word, ProductName etc..
Word2vec: being the algorithm of *** company exploitation, by unsupervised training, by word become a several hundred dimensions to Amount, this vector can capture the semantic dependency between word.Also term vector or word is made to be embedded in.
Tensorflow:Tensorflow is the deep learning platform of *** open source, provides interface abundant, mostly flat Platform (CPU, GPU, HADOOP) and distributed support, visual control.
Skip-gram:Google is used to the training Wordvec in big data and uses method, it predicts week by current word The word enclosed obtains training objective function.
LSTM:LSTM (Long Short-Term Memory) shot and long term memory network, is a kind of time recurrent neural net Network is suitable for being spaced and postpone relatively long critical event in processing and predicted time sequence.It is by " Memory-Gate " and " forgets Note door " controls the going or staying of historical information, efficiently solves the problems, such as the long Route Dependence of conventional recycle neural network.
CRF:CRF (Conditional Random Field) condition random field is that natural language processing field was normal in recent years One of algorithm is usually used in syntactic analysis, name Entity recognition, part-of-speech tagging etc..CRF is using Markov Chain as implicit The probability metastasis model of variable implies variable by Observable condition discrimination, belongs to discrimination model.
Semi-supervised learning: semi-supervised learning (Semi-Supervised Learning, SSL) is pattern-recognition and engineering The Important Problems for practising area research, are a kind of learning methods that supervised learning is combined with unsupervised learning.Semi-supervised learning makes With a large amount of Unlabeled data, and flag data is used simultaneously, Lai Jinhang pattern-recognition work.
Custom Dictionaries: user's Custom Dictionaries are when carrying out NER extraction, and user wishes the special reality extracted Body, in such a way that dictionary is set, it is ensured that it can be extracted.
The present embodiment provides a kind of name entity recognition methods of semi-supervised learning combination Custom Dictionaries, including walk as follows It is rapid:
S1, with unlabeled data pre-training Bi-LSTM language model;
It is had the advantage that using the Bi-LSTM language model of pre-training
1) demand of mark corpus is reduced, the main function of language model is exactly the automatic extraction of feature, using not marking It infuses data and carries out pre-training, obtain the semantic information of each character in advance.
2) training time for reducing model is reduced and is trained with labeled data due to the pre-training carried out in advance Time.
The present invention carries out the training of language model using Bi-LSTM model, is the method for unsupervised learning, is not required to very important person Work marks corpus can training pattern.The network structure of model is as shown in Figure 2.
Bi-LSTM (two-way LSTM) can provide the semantic information of character in conjunction with the information of context, by it is preceding to A LSTM semantic vector available for each character, then again by backward LSTM to another it is semantic to Amount.Two semantic vectors are spliced to obtain final output in output layer.Since the training of this language model is unsupervised Form, therefore the data volume of demand is the bigger the better.
From figure 2 it can be seen that the Bi-LSTM used in the present embodiment, forward and backward and without parameter sharing.Two Layer LSTM is trained using different parameters, that is to say, that two layers of LSTM is independent.
S2, use word2vec model by each character vector at Embedding layers.
In the present embodiment, word vector model is specifically obtained using the training of Skip-gram method.
Each word can be become the vector in a lower dimensional space by word2vec model, usual several hundred dimensions.Such character Between semantic dependency can be with the distance of vector come approximate description.Relative to common term vector, the vector based on character Change technology can bring following advantage:
1) more fine-grained character feature can be characterized;
2) since character quantity is much smaller than word quantity, obtained model occupied space is minimum, greatly improves model and adds Carry speed;
3) over time, neologisms can continue to bring out, and the term vector model trained before will appear increasingly tighter The feature hit rate downslide problem of weight, and the vector based on character then effectively prevents this problem, comes because being continuously created every year Fresh character it is relatively seldom.
Therefore the present embodiment selects the vectorization technology based on character.
The word2vec model that the present embodiment uses is unsupervised learning method, that is, does not need manually to mark corpus Training pattern, common are two kinds of training methods is CBOW and Skip-gram, as shown in Figure 3-4.
CBOW is the word of based on context pre- measured center, according to the character w (t-2), w (t-1), w (t+ around character w (t) 1), w (t+2) is predicted, the vector of these words is connected, can be sufficiently reserved contextual information in this way, as shown in Figure 3.Skip- Gram method is exactly the opposite, removes word w (t-2), w (t-1), w (t+1), the w (t+2) around prediction using w (t), as shown in Figure 4. Under the conditions of big data quantity, it is suitble to use Skip-gram method.
As shown in figure 5, in the present embodiment, using the specific steps of Skip-gram method training pattern are as follows:
(1) collecting relevant balanced corpus first, (because to do unsupervised learning, data volume is the bigger the better, without mark Note), these corpus cover most of data type of the scene mainly for corresponding application field as far as possible;
(2) corpus data collected for step (1) pre-processes, including filters out junk data, stops low-frequency word and nothing Meaning symbol, and be organized into the format of training data, that is, it indicates to output and input, obtains training data;
(3) training data is given to Skip-gram model, training obtains word vector model.
S3, the sequence labelling is trained as sequence labelling model, and using labeled data using two layers of two-way LSTM Model;
In the present embodiment, training data is labeled using BIO mark method.Such as:
Label B-PER then represents the beginning of name, and I-ORG represents the centre of institution term, and O then represents other.
The sequence labelling model of the present embodiment uses two layers of two-way LSTM, due to being carried out using a small amount of labeled data Training, it is contemplated that by the complexity of increase model come better fitting data.While in order to reduce to labeled data magnitude Demand, the present embodiment introduces the language model vector of pre-training between two layers of two-way LSTM of sequence labelling model, has The model of body is as shown in Figure 6.
Specifically, in the training process of sequence labelling model, by the defeated of the two-way LSTM of the first layer of sequence labelling model The output work of outgoing vector and Bi-LSTM language model splices, then the vector of splicing is used as to sequence after a full articulamentum The input of the two-way LSTM of the second layer of column marking model.
From the point of view of concrete implementation, the calculating process of the two-way LSTM of first layer of sequence labelling model, forward direction are initially entered The output of LSTM and backward LSTM, forward direction LSTM are hft, the output of backward LSTM is hbt, after the two is spliced, obtain ht1= [hft,hbt], wherein forward direction exports hftHistory context information is characterized, and exports h backwardbtThen characterize following context Information.Then by the output h of Bi-LSTM language modellmAfter output with the two-way LSTM of first layer is spliced, h is obtainedt= [hlm,ht1].Later after a full articulamentum, in the two-way LSTM of the second layer that result is input to sequence labelling model.
Recognition with Recurrent Neural Network (RNN) is widely applied in natural language processing field at present.For arbitrarily inputting text This sequence (x1,x2,…,xn), RNN returns to the output value set (h for being directed to this sequence1,h2,…,hn).Since traditional RNN exists It carries out that gradient disappearance can be led to the problem of during optimization, so that can not remember long text when prediction Record remote semantic information.And LSTM model controls outputting and inputting for historical information using different gate, meanwhile, it is double To LSTM can not only refer to past historical information, can also be with reference to following semantic information.
S4, user's Custom Dictionaries are added;
Emission matrix X can be obtained after two layers of two-way LSTM of sequence labelling model, by CRF layers, passes through maximum Likelihood probability obtains transfer matrix Y, and the probability of emission matrix, the hair after being adjusted then are adjusted according to user's Custom Dictionaries Penetrate matrix X.
S5, maximum probability path in sequence is found out using Veterbi decoding.
Specifically, CRF layers will be input to according to user's Custom Dictionaries emission matrix X adjusted and transfer matrix Y Veterbi decoding obtains sequence labelling, i.e., correctly names Entity recognition result.
CRF floor is exactly to carry out Veterbi decoding to find optimal road as the main effect of the last layer in the present invention Diameter.Condition random field (conditional random fields, abbreviation CRF or CRFs), is a kind of discriminate probabilistic model, It is one kind of random field, is usually used in mark or analytical sequence data, such as natural language text or biological sequence.Condition random field Model both has the advantages that discriminative model, and with production model in view of the transition probability between contextual tagging, with sequence Column form carries out the characteristics of global parameter optimization and decoding, solves other discriminative model (such as maximum entropy Markov moulds Type) the marking bias problem that is difficult to avoid that.
And condition random field then uses a kind of probability graph model, has the energy of expression long-distance dependence and overlapping property feature Power can preferably solve the advantages of marking the problems such as (classification) biases, and all features can carry out global normalization, energy Enough acquire global optimal solution.Here what is mainly used is exactly the prediction algorithm of condition random field: viterbi algorithm (Viterbi Algorithm, a kind of dynamic programming algorithm).
For those skilled in the art, it can be provided various corresponding according to above technical solution and design Change and modification, and all these change and modification, should be construed as being included within the scope of protection of the claims of the present invention.

Claims (4)

1. a kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries, which comprises the steps of:
S1, with unlabeled data pre-training Bi-LSTM language model;
S2, use word vector model by each character vector at Embedding layers;
S3, using two layers of two-way LSTM as sequence labelling model, the sequence labelling model is using labeled data training;
In the training process of sequence labelling model, by the output vector of the two-way LSTM of the first layer of sequence labelling model and step S1 The output work for the Bi-LSTM language model that middle pre-training obtains splices, then by the vector of splicing after a full articulamentum The input of the two-way LSTM of the second layer as sequence labelling model;
S4, user's Custom Dictionaries are added:
Emission matrix X can be obtained after two layers of two-way LSTM of sequence labelling model, by CRF layers, passes through maximum likelihood Probability obtains transfer matrix Y, and the probability of emission matrix, the transmitting square after being adjusted then are adjusted according to user's Custom Dictionaries Battle array X;
S5, maximum probability path in sequence is found out using Veterbi decoding:
CRF layers will be input to according to user's Custom Dictionaries emission matrix X adjusted and transfer matrix Y obtained in step S4 Veterbi decoding, obtain sequence labelling, i.e., correctly name Entity recognition result.
2. a kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries according to claim 1, special Sign is, in step S2, the word vector model is word2vec model.
3. a kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries according to claim 2, special Sign is, in step S2, specifically carries out the training of word vector model using Skip-gram method.
4. a kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries according to claim 3, special Sign is, the specific steps of word vector model training are carried out using Skip-gram method are as follows:
(1) the relevant balanced corpus of application field is collected first;
(2) corpus data collected for step (1) pre-processes, including filters out junk data, stops low-frequency word and meaningless Symbol, and it is organized into the format of training data, obtain training data;
(3) training data is given to Skip-gram model, training obtains word vector model.
CN201910172675.7A 2019-03-07 2019-03-07 Named entity recognition method combining semi-supervised learning with user-defined dictionary Active CN109858041B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910172675.7A CN109858041B (en) 2019-03-07 2019-03-07 Named entity recognition method combining semi-supervised learning with user-defined dictionary

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910172675.7A CN109858041B (en) 2019-03-07 2019-03-07 Named entity recognition method combining semi-supervised learning with user-defined dictionary

Publications (2)

Publication Number Publication Date
CN109858041A true CN109858041A (en) 2019-06-07
CN109858041B CN109858041B (en) 2023-02-17

Family

ID=66900199

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910172675.7A Active CN109858041B (en) 2019-03-07 2019-03-07 Named entity recognition method combining semi-supervised learning with user-defined dictionary

Country Status (1)

Country Link
CN (1) CN109858041B (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110598203A (en) * 2019-07-19 2019-12-20 中国人民解放军国防科技大学 Military imagination document entity information extraction method and device combined with dictionary
CN111062215A (en) * 2019-12-10 2020-04-24 金蝶软件(中国)有限公司 Named entity recognition method and device based on semi-supervised learning training
CN111079418A (en) * 2019-11-06 2020-04-28 科大讯飞股份有限公司 Named body recognition method and device, electronic equipment and storage medium
CN111079437A (en) * 2019-12-20 2020-04-28 深圳前海达闼云端智能科技有限公司 Entity identification method, electronic equipment and storage medium
CN111079405A (en) * 2019-11-29 2020-04-28 微民保险代理有限公司 Text information identification method and device, storage medium and computer equipment
CN111274814A (en) * 2019-12-26 2020-06-12 浙江大学 Novel semi-supervised text entity information extraction method
CN111274817A (en) * 2020-01-16 2020-06-12 北京航空航天大学 Intelligent software cost measurement method based on natural language processing technology
CN111291550A (en) * 2020-01-17 2020-06-16 北方工业大学 Chinese entity extraction method and device
CN111985240A (en) * 2020-08-19 2020-11-24 腾讯云计算(长沙)有限责任公司 Training method of named entity recognition model, named entity recognition method and device
CN112464645A (en) * 2020-10-30 2021-03-09 中国电力科学研究院有限公司 Semi-supervised learning method, system, equipment, storage medium and semantic analysis method
CN113761215A (en) * 2021-03-25 2021-12-07 中科天玑数据科技股份有限公司 Feedback self-learning-based dynamic dictionary base generation method
WO2022022421A1 (en) * 2020-07-29 2022-02-03 北京字节跳动网络技术有限公司 Language representation model system, pre-training method and apparatus, device and medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107133220A (en) * 2017-06-07 2017-09-05 东南大学 Name entity recognition method in a kind of Geography field
CN107797992A (en) * 2017-11-10 2018-03-13 北京百分点信息科技有限公司 Name entity recognition method and device
CN108388560A (en) * 2018-03-17 2018-08-10 北京工业大学 GRU-CRF meeting title recognition methods based on language model
CN108628823A (en) * 2018-03-14 2018-10-09 中山大学 In conjunction with the name entity recognition method of attention mechanism and multitask coordinated training
WO2018218705A1 (en) * 2017-05-27 2018-12-06 中国矿业大学 Method for recognizing network text named entity based on neural network probability disambiguation
CN109284400A (en) * 2018-11-28 2019-01-29 电子科技大学 A kind of name entity recognition method based on Lattice LSTM and language model

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018218705A1 (en) * 2017-05-27 2018-12-06 中国矿业大学 Method for recognizing network text named entity based on neural network probability disambiguation
CN107133220A (en) * 2017-06-07 2017-09-05 东南大学 Name entity recognition method in a kind of Geography field
CN107797992A (en) * 2017-11-10 2018-03-13 北京百分点信息科技有限公司 Name entity recognition method and device
CN108628823A (en) * 2018-03-14 2018-10-09 中山大学 In conjunction with the name entity recognition method of attention mechanism and multitask coordinated training
CN108388560A (en) * 2018-03-17 2018-08-10 北京工业大学 GRU-CRF meeting title recognition methods based on language model
CN109284400A (en) * 2018-11-28 2019-01-29 电子科技大学 A kind of name entity recognition method based on Lattice LSTM and language model

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
CHRISTOPHER CLARK等: "Deep contextualized word representations", 《HTTPS://ARXIV.ORG/PDF/1802.05365.PDF》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110598203A (en) * 2019-07-19 2019-12-20 中国人民解放军国防科技大学 Military imagination document entity information extraction method and device combined with dictionary
CN111079418A (en) * 2019-11-06 2020-04-28 科大讯飞股份有限公司 Named body recognition method and device, electronic equipment and storage medium
CN111079418B (en) * 2019-11-06 2023-12-05 科大讯飞股份有限公司 Named entity recognition method, device, electronic equipment and storage medium
CN111079405A (en) * 2019-11-29 2020-04-28 微民保险代理有限公司 Text information identification method and device, storage medium and computer equipment
CN111062215A (en) * 2019-12-10 2020-04-24 金蝶软件(中国)有限公司 Named entity recognition method and device based on semi-supervised learning training
CN111062215B (en) * 2019-12-10 2024-02-13 金蝶软件(中国)有限公司 Named entity recognition method and device based on semi-supervised learning training
CN111079437A (en) * 2019-12-20 2020-04-28 深圳前海达闼云端智能科技有限公司 Entity identification method, electronic equipment and storage medium
CN111274814A (en) * 2019-12-26 2020-06-12 浙江大学 Novel semi-supervised text entity information extraction method
CN111274814B (en) * 2019-12-26 2021-09-24 浙江大学 Novel semi-supervised text entity information extraction method
CN111274817A (en) * 2020-01-16 2020-06-12 北京航空航天大学 Intelligent software cost measurement method based on natural language processing technology
CN111291550A (en) * 2020-01-17 2020-06-16 北方工业大学 Chinese entity extraction method and device
CN111291550B (en) * 2020-01-17 2021-09-03 北方工业大学 Chinese entity extraction method and device
WO2022022421A1 (en) * 2020-07-29 2022-02-03 北京字节跳动网络技术有限公司 Language representation model system, pre-training method and apparatus, device and medium
CN111985240A (en) * 2020-08-19 2020-11-24 腾讯云计算(长沙)有限责任公司 Training method of named entity recognition model, named entity recognition method and device
CN111985240B (en) * 2020-08-19 2024-02-27 腾讯云计算(长沙)有限责任公司 Named entity recognition model training method, named entity recognition method and named entity recognition device
CN112464645A (en) * 2020-10-30 2021-03-09 中国电力科学研究院有限公司 Semi-supervised learning method, system, equipment, storage medium and semantic analysis method
CN113761215A (en) * 2021-03-25 2021-12-07 中科天玑数据科技股份有限公司 Feedback self-learning-based dynamic dictionary base generation method

Also Published As

Publication number Publication date
CN109858041B (en) 2023-02-17

Similar Documents

Publication Publication Date Title
CN109858041A (en) A kind of name entity recognition method of semi-supervised learning combination Custom Dictionaries
CN110334354B (en) Chinese relation extraction method
CN110287481B (en) Named entity corpus labeling training system
CN108460013B (en) Sequence labeling model and method based on fine-grained word representation model
CN110633409B (en) Automobile news event extraction method integrating rules and deep learning
CN107168945B (en) Bidirectional cyclic neural network fine-grained opinion mining method integrating multiple features
CN110929030B (en) Text abstract and emotion classification combined training method
CN107330032B (en) Implicit discourse relation analysis method based on recurrent neural network
CN111241294B (en) Relationship extraction method of graph convolution network based on dependency analysis and keywords
CN111046179B (en) Text classification method for open network question in specific field
CN110532557B (en) Unsupervised text similarity calculation method
CN109003601A (en) A kind of across language end-to-end speech recognition methods for low-resource Tujia language
CN109871538A (en) A kind of Chinese electronic health record name entity recognition method
CN109871535A (en) A kind of French name entity recognition method based on deep neural network
CN110569508A (en) Method and system for classifying emotional tendencies by fusing part-of-speech and self-attention mechanism
CN111062217B (en) Language information processing method and device, storage medium and electronic equipment
CN110263325A (en) Chinese automatic word-cut
CN109189862A (en) A kind of construction of knowledge base method towards scientific and technological information analysis
CN114492441A (en) BilSTM-BiDAF named entity identification method based on machine reading understanding
CN111222318B (en) Trigger word recognition method based on double-channel bidirectional LSTM-CRF network
CN113761893B (en) Relation extraction method based on mode pre-training
CN110287482A (en) Semi-automation participle corpus labeling training device
Li et al. UD_BBC: Named entity recognition in social network combined BERT-BiLSTM-CRF with active learning
CN111914556A (en) Emotion guiding method and system based on emotion semantic transfer map
CN114490991A (en) Dialog structure perception dialog method and system based on fine-grained local information enhancement

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 100081 No.101, 1st floor, building 14, 27 Jiancai Chengzhong Road, Haidian District, Beijing

Applicant after: Beijing PERCENT Technology Group Co.,Ltd.

Address before: 100081 16 / F, block a, Beichen Century Center, building 2, courtyard 8, Beichen West Road, Chaoyang District, Beijing

Applicant before: BEIJING BAIFENDIAN INFORMATION SCIENCE & TECHNOLOGY Co.,Ltd.

GR01 Patent grant
GR01 Patent grant