CN111160042B - Text semantic analysis method and device - Google Patents

Text semantic analysis method and device Download PDF

Info

Publication number
CN111160042B
CN111160042B CN201911406192.5A CN201911406192A CN111160042B CN 111160042 B CN111160042 B CN 111160042B CN 201911406192 A CN201911406192 A CN 201911406192A CN 111160042 B CN111160042 B CN 111160042B
Authority
CN
China
Prior art keywords
vector matrix
text
decoder
generating
context
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201911406192.5A
Other languages
Chinese (zh)
Other versions
CN111160042A (en
Inventor
马雄
杨旭川
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing Juexiao Technology Co ltd
Original Assignee
Chongqing Juexiao Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing Juexiao Technology Co ltd filed Critical Chongqing Juexiao Technology Co ltd
Priority to CN201911406192.5A priority Critical patent/CN111160042B/en
Publication of CN111160042A publication Critical patent/CN111160042A/en
Application granted granted Critical
Publication of CN111160042B publication Critical patent/CN111160042B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/16Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Pure & Applied Mathematics (AREA)
  • Mathematical Analysis (AREA)
  • Mathematical Optimization (AREA)
  • Computational Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computing Systems (AREA)
  • Algebra (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Machine Translation (AREA)

Abstract

The embodiment of the application provides a text semantic analysis method and a text semantic analysis device, which are used for solving the problem of low accuracy of the semantic analysis method in the prior art. The method comprises the following steps: providing and training an analytical model comprising an encoder and a decoder; encoding by the encoder based on the text to be parsed, and generating a first vector matrix; the text to be analyzed comprises a short text and a long text; generating, by the decoder, a context vector matrix based on the first vector matrix, the context vector matrix characterizing contributions to itself of words other than itself; determining a boundary vector matrix based on the context vector matrix, decoding the boundary vector matrix by the decoder, determining whether the long text contains the short text.

Description

Text semantic analysis method and device
Technical Field
The present disclosure relates to the field of text information processing technologies, and in particular, to a text semantic parsing method and apparatus.
Background
With the development of computer technology, many application scenarios are trying to participate in computers, so as to improve efficiency. For example, scoring subjective questions of an examination; also for example, audits for certain text, and the like. In these application scenarios, it is often necessary to determine the similarity between two texts, for example, according to the determined standard answer or the key sentence of the standard answer, to determine whether the semantic meaning of the subjective question answer contains the key sentence of the standard answer.
In the prior art, the similarity between two texts is generally determined based on word senses represented by each word in the two texts. For a sentence or a paragraph of sentences, much other information is included in addition to being composed of individual words. Even if the same phrase constitutes a sentence, other information is different, which causes the meaning expressed by the sentence to be different, and thus the final judgment result is affected. Therefore, how to score according to more comprehensive text information and obtain more accurate scoring results is a problem to be solved.
Disclosure of Invention
The purpose of the application is to provide a text semantic analysis method and a text semantic analysis device, so as to solve the problem of low accuracy of the semantic analysis method in the prior art.
In a first aspect of the present application, a text semantic parsing method is provided, including:
providing and training an analytical model comprising an encoder and a decoder;
encoding by the encoder based on the text to be parsed, and generating a first vector matrix; the text to be analyzed comprises a short text and a long text;
generating, by the decoder, a context vector matrix based on the first vector matrix, the context vector matrix characterizing contributions to itself of words other than itself;
determining a boundary vector matrix based on the context vector matrix, decoding the boundary vector matrix by the decoder, determining whether the long text contains the short text.
In a possible implementation manner, the encoding, by the encoder, based on the text to be parsed, generates a first vector matrix, including:
the encoder is used for inputting the short text and the long text according to a set format by taking a Bert model as the encoder, wherein the set format is [ classification label ] short text [ text segmenter ] long text [ text segmenter ];
and encoding the text to be analyzed through a Bert model to generate the first vector matrix, wherein the number of lines of the first vector matrix is the sum of the lengths of the contents input by adopting the set format.
In a possible implementation manner, the generating, by the decoder, a context vector matrix based on the first vector matrix includes:
generating, by the decoder, a second vector matrix from the long text by determining cues of the short text based on the first vector matrix;
the context vector matrix is generated by the decoder based on the second vector matrix.
In a possible implementation manner, the generating the second vector matrix and the context vector matrix specifically includes:
generating a second vector matrix based on the first vector matrix and DGCNN model;
and generating the context vector matrix based on the second vector matrix and a Self-attribute model.
In a possible implementation manner, the generating a second vector matrix based on the first vector matrix and DGCNN model includes:
generating a first intermediate vector matrix by performing one-dimensional expansion convolution on the first vector matrix;
performing one-dimensional expansion convolution on the first vector matrix, and activating the first vector matrix by using an activation function to generate a second intermediate vector matrix;
the second vector matrix is generated by using the second intermediate vector matrix as a weight based on the first vector matrix, the first intermediate vector matrix, and the second intermediate vector matrix.
In a second aspect of the present application, there is provided a text semantic parsing apparatus, comprising an encoder and a decoder,
the encoder is used for encoding the text to be analyzed and generating a first vector matrix; the text to be analyzed comprises a short text and a long text;
the decoder is configured to generate a context vector matrix based on the first vector matrix; and determining a boundary vector matrix based on the context vector matrix, decoding the boundary vector matrix, and determining whether the long text contains the short text.
In a possible implementation manner, the encoder is specifically configured to:
the encoder is used for inputting the short text and the long text according to a set format by taking a Bert model as the encoder, wherein the set format is [ classification label ] short text [ text segmenter ] long text [ text segmenter ];
and encoding the text to be analyzed through a Bert model to generate the first vector matrix, wherein the number of lines of the first vector matrix is the sum of the lengths of the contents input by adopting the set format.
In a possible implementation manner, the decoder generates, based on the first vector matrix, a context vector matrix by the decoder, including:
generating, by the decoder, a second vector matrix from the long text by determining cues of the short text based on the first vector matrix;
the context vector matrix is generated by the decoder based on the first vector matrix and the second vector matrix.
In a possible implementation manner, the decoder generates a second vector matrix and the context vector matrix, specifically including:
generating a second vector matrix based on the first vector matrix and DGCNN model;
the context vector matrix is generated based on the first vector matrix, the second vector matrix, and a Self-intent model.
In a possible implementation manner, the decoder generates a second vector matrix based on the first vector matrix and DGCNN model, including:
generating a first intermediate vector matrix by performing one-dimensional expansion convolution on the first vector matrix;
performing one-dimensional expansion convolution on the first vector matrix, and activating the first vector matrix by using an activation function to generate a second intermediate vector matrix;
the second vector matrix is generated by using the second intermediate vector matrix as a weight based on the first vector matrix, the first intermediate vector matrix, and the second intermediate vector matrix.
In the embodiment of the application, the provided parsing model comprises an encoder and a decoder, after training the parsing model, the text to be parsed can be encoded to obtain a first vector matrix, the context vector matrix is generated by subsequent decoding based on the first vector matrix, the boundary vector matrix is determined according to the context vector matrix, and whether the long text contains the short text is determined by decoding the boundary vector matrix through the decoder. Therefore, the semantic text of the short text is found in the long text, the accurate analysis of subjective questions is further realized, and the accuracy of analysis is improved.
Drawings
FIG. 1 is a schematic diagram of an application architecture according to an embodiment of the present application;
fig. 2 is a flowchart of a text semantic parsing method provided in an embodiment of the present application;
FIG. 3 is a flowchart of encoding text to be parsed according to an embodiment of the present application;
FIG. 4 is a flow chart of generating a context vector matrix provided by an embodiment of the present application;
FIG. 5 is a flowchart illustrating a method for generating a context vector matrix according to an embodiment of the present disclosure;
FIG. 6 is a specific flow chart for generating a second vector matrix provided by an embodiment of the present application;
fig. 7 is a schematic structural diagram of a text semantic parsing device according to an embodiment of the present application.
Detailed Description
For the purposes of making the objects, technical solutions and advantages of the embodiments of the present application more clear, the technical solutions of the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is apparent that the described embodiments are some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by one of ordinary skill in the art without undue burden from the present disclosure, are within the scope of the present application based on the embodiments herein.
It should be noted that the same or similar reference numerals denote the same or similar elements or elements having the same or similar functions throughout. The embodiments described below by referring to the drawings are exemplary only for the purpose of explaining the present application and are not to be construed as limiting the present application.
The architecture applied in the present application may be as shown in fig. 1, providing a parsing model comprising an encoder and a decoder. The encoder and decoder may be designed as different modules of the same computing device or may be separate modules. It is apparent that the encoder and decoder can operate a text semantic parsing method provided in the embodiments of the present application, as shown in fig. 2, including the following steps 201 to 204:
training an analytical model comprising an encoder and a decoder 201;
202, encoding by an encoder based on a text to be parsed, and generating a first vector matrix; the text to be parsed comprises short text and long text.
203, generating, by the decoder, a context vector matrix based on the first vector matrix, the context vector matrix characterizing contributions of words other than itself to itself.
204, determining a boundary vector matrix based on the context vector matrix, decoding the boundary vector matrix by a decoder, and determining whether the long text contains the short text.
In the embodiment of the application, the provided parsing model comprises an encoder and a decoder, after training the parsing model, the text to be parsed can be encoded to obtain a first vector matrix, the context vector matrix is generated by subsequent decoding based on the first vector matrix, the boundary vector matrix is determined according to the context vector matrix, and whether the long text contains the short text is determined by decoding the boundary vector matrix through the decoder. Therefore, the semantic text of the short text is found in the long text, the accurate analysis of subjective questions is further realized, and the accuracy of analysis is improved.
It should be noted that, step 202 and step 203 may be used as part of training the analytical model, and after step 203, the training of maximizing the probability may be performed, so that step 201 may be implemented.
It should be noted that the term "in this application may refer to a chinese word or a foreign word, or a combination of words specified in some possible embodiments.
In a possible implementation, as shown in fig. 3, step 202 includes:
301, using the Bert model as an encoder, an encoder for inputting short text and long text in a set format. Wherein the set format may be [ class label ] short text [ text separator ] long text [ text separator ].
302, encoding text to be parsed through a Bert model, and generating a first vector matrix. The number of rows of the first vector matrix is the sum of the lengths of the contents input in the set format.
In a possible implementation, as shown in fig. 4, step 203 includes:
a second vector matrix is generated 401 by a decoder determining cues of short text from long text based on the first vector matrix.
A context vector matrix is generated by a decoder based on the second vector matrix 402.
Based on the representation of fig. 4, as shown in fig. 5, in some possible examples, generating the second vector matrix specifically includes: 501, generating a second vector matrix based on the first vector matrix and the DGCNN model. In some possible examples, generating the context vector matrix specifically includes: a context vector matrix is generated 502 based on the second vector matrix and the Self-intent model.
For step 501, in some possible examples, a method as shown in fig. 6 may be used, including:
601, generating a first intermediate vector matrix by performing one-dimensional expansion convolution on the first vector matrix;
602, performing one-dimensional expansion convolution on the first vector matrix, and activating with an activation function to generate a second intermediate vector matrix;
603, generating a second vector matrix by using the second intermediate vector matrix as weights based on the first vector matrix, the first intermediate vector matrix and the second intermediate vector matrix.
From the above embodiments, the embodiments of the present application provide a specific architecture diagram of an analysis model, as shown in fig. 7:
the parsing model includes a Bert model as an encoder 1 of the parsing model, a DGCNN model and a Self-section model as part of a decoder 2. For a clearer understanding of the present solution, based on the specific architecture diagram described above, the following steps are described:
step one, inputting a text to be analyzed into a Bert model, outputting a first vector matrix, and marking the first vector matrix as y; as in equation 1:
y=Transformer encoder (x,z); (1)
in this application, the Bert model plays a role here in encoding text to be parsed into a vector matrix, and Bert is composed of 12 layers transformer encoder. The text to be analyzed is spliced by a long text z and a short text x, the length after combination is L, L is the sum of the lengths of the long text z, the short text x, the classification labels and the segmenters, and the specific format is [ CLS ] short text [ SEP ] long text [ SEP ], wherein: CLS is a class label in Bert, [ SEP ] is a text separator in Bert. In some examples, the first vector matrix may be a l×768-dimensional vector matrix.
And step two, inputting the first vector matrix into the DGCNN model, and outputting a second vector matrix. The main role of the DGCNN model here is to set all content information in the long text z that is not related to the short text x to null. Here there are 4 different DGCNNs, the difference is in the magnitude of the different expansion convolution rates, which are 1,2,4,8, so there are 4 DGCNNs, the output of the last DGCNN is taken as the input of the next DGCNN, as shown in equation 2:
Figure BDA0002348702160000081
wherein g is DGCNN model output, conv1D 1 And Conv1D 2 Is a one-dimensional expansion convolution functionThe number, σ, is the activation function,
Figure BDA0002348702160000091
is a tensor product operator. Clearly Conv1D 1 The activation function is not applied and,
it is only a linear transformation, so Conv1D 1 (y) -y can be combined together, equivalent to a single Conv1D 1
Taking the first vector matrix as l×768-dimensional vector matrix as an example, the description will be given. Specific:
the first vector matrix y is subjected to one-dimensional expansion convolution, the expansion rate is 1, the convolution window is 3 x 768, the number of convolution kernels is 768, the function of the first vector matrix y is linear transformation, and a first intermediate vector matrix is obtained, and the first intermediate vector matrix is an L x 768-dimensional vector and is marked as g1.
The first vector matrix y is subjected to one-dimensional expansion convolution, the expansion rate is 1, the convolution window is 3 x 768, and the convolution window is activated by using a sigmoid function (i.e. sigma function), wherein the function is to convert the vector into a probability, so as to obtain a second intermediate vector matrix, and the second intermediate vector matrix is an L x 768-dimensional vector matrix and is marked as g2.
Based on the first vector matrix, the first intermediate vector matrix, and the second intermediate vector matrix, a second vector matrix is generated by using the second intermediate vector matrix as a weight, and the second vector matrix is an l×768-dimensional vector matrix, denoted as g3. As in equation 3:
g3=y*(1-g2)+g1*g2; (3)
i.e. the information is represented to pass directly with a probability of 1-g2 and is transformed with a probability of g2.
And thirdly, inputting the second vector matrix g3 into a Self-attribute model to generate a context vector matrix h. The Self-intent model plays a role here to make the information related to the short text more condensed out of the long text, and can be used as follows with equations 4 and 5.
Figure BDA0002348702160000101
h=Self Attention(g); (5)
Wherein Q, K and V are L×768-dimensional vector matrix.
The present application also provides a pointer to complete final decoding, in which, in step four, a context vector matrix h is input into the pointer, a boundary vector matrix is generated through the context vector matrix h, the boundary vector matrix is decoded, the upper boundary and the lower boundary of a sentence which can prove a short text are searched, and whether the long text contains the short text is determined. The method specifically comprises the following steps:
searching a starting position, performing linear transformation on the first vector matrix y, and outputting a vector matrix b1 of L1 x 1; searching an end position, carrying out linear transformation on the first vector matrix y, and outputting a vector matrix b2 of L1 x 1; taking the position p1 of the maximum value of b1; taking the position p2 of the maximum value of b2; the content between p1 and p2 is a sentence which can prove short text. If p1 and p2 are both in the [ CLS ] position, then this long text is indicated as having no short text; otherwise, it indicates that there is short text in this long text.
In a second aspect of the present application, as shown in fig. 7, there is provided a text semantic parsing apparatus, including an encoder 1 and a decoder 2,
the encoder 1 is used for encoding the text to be analyzed to generate a first vector matrix; the text to be analyzed comprises a short text and a long text;
a decoder 2 for generating a context vector matrix based on the first vector matrix; and determining a boundary vector matrix based on the context vector matrix, decoding the boundary vector matrix, and determining whether the long text contains the short text.
In a possible implementation, the encoder 1 is specifically configured to:
taking the Bert model as an encoder 1, and inputting short texts and long texts according to a set format, wherein the set format is [ classification labels ] short texts [ text separators ] long texts [ text separators ];
and encoding the text to be analyzed through the Bert model to generate a first vector matrix, wherein the number of lines of the first vector matrix is the sum of the lengths of the contents input in a set format.
In a possible implementation, the decoder 2 generates, by the decoder, a context vector matrix based on the first vector matrix, including:
generating a second vector matrix from the long text by determining cues of the short text by a decoder based on the first vector matrix;
a context vector matrix is generated by a decoder based on the first vector matrix and the second vector matrix.
In a possible implementation manner, the decoder 2 generates a second vector matrix and a context vector matrix, specifically including:
generating a second vector matrix based on the first vector matrix and the DGCNN model;
a context vector matrix is generated based on the first vector matrix, the second vector matrix, and the Self-intent model.
In a possible implementation, the decoder generates a second vector matrix based on the first vector matrix and the DGCNN model, including:
generating a first intermediate vector matrix by performing one-dimensional expansion convolution on the first vector matrix;
then, performing one-dimensional expansion convolution on the first vector matrix, and activating the first vector matrix by using an activation function to generate a second intermediate vector matrix;
based on the first vector matrix, the first intermediate vector matrix, and the second intermediate vector matrix, the second vector matrix is generated by using the second intermediate vector matrix as a weight.
In the embodiment of the application, the provided parsing model comprises an encoder and a decoder, after training the parsing model, the text to be parsed can be encoded to obtain a first vector matrix, the context vector matrix is generated by subsequent decoding based on the first vector matrix, the boundary vector matrix is determined according to the context vector matrix, and whether the long text contains the short text is determined by decoding the boundary vector matrix through the decoder. The method has the advantages that the Bert model is used as a powerful encoder, the DGCNN door is used for mechanical design of a text inclusion model, and the Self-intent model is combined, so that the semantic text of the short text is found in the long text, further, the accurate analysis of subjective questions is realized, and the analysis accuracy is improved.
In the present embodiments, for a software implementation, the techniques described herein may be implemented by means of units that perform the functions described herein. The software codes may be stored in a memory and executed by a processor. The memory may be implemented within the processor or external to the processor.
Those of ordinary skill in the art will appreciate that the various illustrative elements and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
It will be clear to those skilled in the art that, for convenience and brevity of description, specific working procedures of the above-described systems, apparatuses and units may refer to corresponding procedures in the foregoing method embodiments, and are not repeated herein.
The above-described apparatus embodiments are merely illustrative, for example, the division of the units is merely a logical function division, and there may be additional divisions when actually implemented, for example, multiple units or components may be combined or integrated into another system, or some features may be omitted or not performed. Alternatively, the coupling or direct coupling or communication connection shown or discussed with each other may be an indirect coupling or communication connection via some interfaces, devices or units, which may be in electrical, mechanical or other form.
The units described as separate units may or may not be physically separate, and units shown as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution of this embodiment.
In addition, each functional unit in each embodiment of the present application may be integrated in one processing unit, or each unit may exist alone physically, or two or more units may be integrated in one unit.
The functions, if implemented in the form of software functional units and sold or used as a stand-alone product, may be stored in a computer-readable storage medium. Based on such understanding, the technical solutions of the embodiments of the present application may be essentially or, what contributes to the prior art, or part of the technical solutions, may be embodied in the form of a software product stored in a storage medium, including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to perform all or part of the steps of the methods described in the embodiments of the present application. And the aforementioned storage medium includes: a usb disk, a removable hard disk, a ROM, a RAM, a magnetic disk, or an optical disk, etc.
Although embodiments of the present application have been described in connection with the accompanying drawings, various modifications and variations may be made by those skilled in the art without departing from the spirit and scope of the application, and such modifications and variations are within the scope defined by the appended claims.

Claims (8)

1. A text semantic parsing method, comprising:
providing and training an analytical model comprising an encoder and a decoder;
based on a text to be analyzed, using a Bert model as the encoder, and inputting a short text and a long text included in the text to be analyzed into the encoder according to a set format, wherein the set format is [ classification label ] short text [ text segmenter ] long text [ text segmenter ]; encoding the text to be analyzed through the Bert model to generate a first vector matrix, wherein the number of lines of the first vector matrix is the sum of the lengths of the contents input by adopting the set format;
generating, by the decoder, a context vector matrix based on the first vector matrix, the context vector matrix characterizing contributions to itself of words other than itself;
determining a boundary vector matrix based on the context vector matrix, decoding the boundary vector matrix by the decoder, determining whether the long text contains the short text.
2. The method of claim 1, wherein the generating, by the decoder, the context vector matrix based on the first vector matrix comprises:
generating, by the decoder, a second vector matrix from the long text by determining cues of the short text based on the first vector matrix;
the context vector matrix is generated by the decoder based on the second vector matrix.
3. The method of claim 2, wherein the generating the second vector matrix and the context vector matrix specifically comprises:
generating the second vector matrix based on the first vector matrix and DGCNN model;
and generating the context vector matrix based on the second vector matrix and a Self-attribute model.
4. The method of claim 3, wherein the generating the second vector matrix based on the first vector matrix and the DGCNN model comprises:
generating a first intermediate vector matrix by performing one-dimensional expansion convolution on the first vector matrix;
performing one-dimensional expansion convolution on the first vector matrix, and activating the first vector matrix by using an activation function to generate a second intermediate vector matrix;
the second vector matrix is generated by using the second intermediate vector matrix as a weight based on the first vector matrix, the first intermediate vector matrix, and the second intermediate vector matrix.
5. A text semantic parsing apparatus is characterized by comprising an encoder and a decoder,
the encoder is used for inputting short texts and long texts included in the text to be analyzed into the encoder according to a set format by taking a Bert model as the encoder based on the text to be analyzed, wherein the set format is [ classification labels ] short texts [ text separators ] long texts [ text separators ]; encoding the text to be analyzed through the Bert model to generate a first vector matrix, wherein the number of lines of the first vector matrix is the sum of the lengths of the contents input by adopting the set format;
the decoder is configured to generate a context vector matrix based on the first vector matrix; and determining a boundary vector matrix based on the context vector matrix, decoding the boundary vector matrix, and determining whether the long text contains the short text.
6. The apparatus of claim 5, wherein the decoder generating, by the decoder, the context vector matrix based on the first vector matrix comprises:
generating, by the decoder, a second vector matrix from the long text by determining cues of the short text based on the first vector matrix;
the context vector matrix is generated by the decoder based on the first vector matrix and the second vector matrix.
7. The apparatus of claim 6, wherein the decoder generates the second vector matrix and the context vector matrix, comprising:
generating the second vector matrix based on the first vector matrix and DGCNN model;
the context vector matrix is generated based on the first vector matrix, the second vector matrix, and a Self-intent model.
8. The apparatus of claim 7, wherein the decoder generates the second vector matrix based on the first vector matrix and the DGCNN model, comprising:
generating a first intermediate vector matrix by performing one-dimensional expansion convolution on the first vector matrix;
performing one-dimensional expansion convolution on the first vector matrix, and activating the first vector matrix by using an activation function to generate a second intermediate vector matrix;
the second vector matrix is generated by using the second intermediate vector matrix as a weight based on the first vector matrix, the first intermediate vector matrix, and the second intermediate vector matrix.
CN201911406192.5A 2019-12-31 2019-12-31 Text semantic analysis method and device Active CN111160042B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911406192.5A CN111160042B (en) 2019-12-31 2019-12-31 Text semantic analysis method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911406192.5A CN111160042B (en) 2019-12-31 2019-12-31 Text semantic analysis method and device

Publications (2)

Publication Number Publication Date
CN111160042A CN111160042A (en) 2020-05-15
CN111160042B true CN111160042B (en) 2023-04-28

Family

ID=70559873

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911406192.5A Active CN111160042B (en) 2019-12-31 2019-12-31 Text semantic analysis method and device

Country Status (1)

Country Link
CN (1) CN111160042B (en)

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103257957A (en) * 2012-02-15 2013-08-21 深圳市腾讯计算机***有限公司 Chinese word segmentation based text similarity identifying method and device
CN104834747A (en) * 2015-05-25 2015-08-12 中国科学院自动化研究所 Short text classification method based on convolution neutral network
CN106919556A (en) * 2017-02-21 2017-07-04 重庆邮电大学 A kind of natural language semanteme deep analysis algorithm of use sparse coding
CN107656990A (en) * 2017-09-14 2018-02-02 中山大学 A kind of file classification method based on two aspect characteristic informations of word and word
CN109446333A (en) * 2019-01-16 2019-03-08 深兰人工智能芯片研究院(江苏)有限公司 A kind of method that realizing Chinese Text Categorization and relevant device
CN109684626A (en) * 2018-11-16 2019-04-26 深思考人工智能机器人科技(北京)有限公司 Method for recognizing semantics, model, storage medium and device
CN110019793A (en) * 2017-10-27 2019-07-16 阿里巴巴集团控股有限公司 A kind of text semantic coding method and device
CN110347793A (en) * 2019-06-28 2019-10-18 北京牡丹电子集团有限责任公司宁安智慧工程中心 A kind of semantic analysis method and device of Chinese
CN110569500A (en) * 2019-07-23 2019-12-13 平安国际智慧城市科技股份有限公司 Text semantic recognition method and device, computer equipment and storage medium

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101303692B (en) * 2008-06-19 2012-08-29 徐文和 All-purpose numeral semantic library for translation of mechanical language
JP6583686B2 (en) * 2015-06-17 2019-10-02 パナソニックIpマネジメント株式会社 Semantic information generation method, semantic information generation device, and program
CN108287858B (en) * 2017-03-02 2021-08-10 腾讯科技(深圳)有限公司 Semantic extraction method and device for natural language

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103257957A (en) * 2012-02-15 2013-08-21 深圳市腾讯计算机***有限公司 Chinese word segmentation based text similarity identifying method and device
CN104834747A (en) * 2015-05-25 2015-08-12 中国科学院自动化研究所 Short text classification method based on convolution neutral network
CN106919556A (en) * 2017-02-21 2017-07-04 重庆邮电大学 A kind of natural language semanteme deep analysis algorithm of use sparse coding
CN107656990A (en) * 2017-09-14 2018-02-02 中山大学 A kind of file classification method based on two aspect characteristic informations of word and word
CN110019793A (en) * 2017-10-27 2019-07-16 阿里巴巴集团控股有限公司 A kind of text semantic coding method and device
CN109684626A (en) * 2018-11-16 2019-04-26 深思考人工智能机器人科技(北京)有限公司 Method for recognizing semantics, model, storage medium and device
CN109446333A (en) * 2019-01-16 2019-03-08 深兰人工智能芯片研究院(江苏)有限公司 A kind of method that realizing Chinese Text Categorization and relevant device
CN110347793A (en) * 2019-06-28 2019-10-18 北京牡丹电子集团有限责任公司宁安智慧工程中心 A kind of semantic analysis method and device of Chinese
CN110569500A (en) * 2019-07-23 2019-12-13 平安国际智慧城市科技股份有限公司 Text semantic recognition method and device, computer equipment and storage medium

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
"Approach of Text Search Based on Semantic Parsing Model ".《Fourth International Conference on Fuzzy Systems and Knowledge Discovery (FSKD 2007)》.2007,全文. *
杨俊峰."基于word2vec和CNN的短文本聚类研究".信息与电脑.2019,全文. *
杨越 ; 黄瑞章 ; 魏琴 ; 陈艳平 ; 秦永彬 ; .基于上下文语义的新闻人名纠错方法.电子科技大学学报.2019,(第06期),全文. *
黄栋 ; 徐博 ; 许侃 ; 林鸿飞 ; 杨志豪 ; .基于词向量和EMD距离的短文本聚类.山东大学学报(理学版).2017,(第07期),全文. *

Also Published As

Publication number Publication date
CN111160042A (en) 2020-05-15

Similar Documents

Publication Publication Date Title
US11106714B2 (en) Summary generating apparatus, summary generating method and computer program
CN108717406B (en) Text emotion analysis method and device and storage medium
KR102401942B1 (en) Method and apparatus for evaluating translation quality
CN111931517B (en) Text translation method, device, electronic equipment and storage medium
CN112417102B (en) Voice query method, device, server and readable storage medium
CN111324743A (en) Text relation extraction method and device, computer equipment and storage medium
CN112528637B (en) Text processing model training method, device, computer equipment and storage medium
CN109740158B (en) Text semantic parsing method and device
CN110795552A (en) Training sample generation method and device, electronic equipment and storage medium
CN111401084A (en) Method and device for machine translation and computer readable storage medium
CN113268576B (en) Deep learning-based department semantic information extraction method and device
CN110427619B (en) Chinese text automatic proofreading method based on multi-channel fusion and reordering
CN111985243B (en) Emotion model training method, emotion analysis device and storage medium
CN111930792A (en) Data resource labeling method and device, storage medium and electronic equipment
CN113158656B (en) Ironic content recognition method, ironic content recognition device, electronic device, and storage medium
CN112016320A (en) English punctuation adding method, system and equipment based on data enhancement
CN113255331B (en) Text error correction method, device and storage medium
CN116628186B (en) Text abstract generation method and system
CN111159405B (en) Irony detection method based on background knowledge
CN113761883A (en) Text information identification method and device, electronic equipment and storage medium
CN116050425A (en) Method for establishing pre-training language model, text prediction method and device
CN111160042B (en) Text semantic analysis method and device
CN116561271A (en) Question and answer processing method and device
KR102422844B1 (en) Method of managing language risk of video content based on artificial intelligence
CN114925175A (en) Abstract generation method and device based on artificial intelligence, computer equipment and medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: 401120 office 21, building 1, Xizheng international academic exchange center, No. 209, Baosheng Avenue, Huixing street, Yubei District, Chongqing

Applicant after: Chongqing juexiao Technology Co.,Ltd.

Address before: 401120 office 21, building 1, Xizheng international academic exchange center, No. 209, Baosheng Avenue, Huixing street, Yubei District, Chongqing

Applicant before: Chongqing juexing Education Technology Co.,Ltd.

GR01 Patent grant
GR01 Patent grant