CN109145107A - Subject distillation method, apparatus, medium and equipment based on convolutional neural networks - Google Patents

Subject distillation method, apparatus, medium and equipment based on convolutional neural networks Download PDF

Info

Publication number
CN109145107A
CN109145107A CN201811133725.2A CN201811133725A CN109145107A CN 109145107 A CN109145107 A CN 109145107A CN 201811133725 A CN201811133725 A CN 201811133725A CN 109145107 A CN109145107 A CN 109145107A
Authority
CN
China
Prior art keywords
output
term vector
region unit
hidden layer
text
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201811133725.2A
Other languages
Chinese (zh)
Other versions
CN109145107B (en
Inventor
金戈
徐亮
肖京
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ping An Technology Shenzhen Co Ltd
Original Assignee
Ping An Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ping An Technology Shenzhen Co Ltd filed Critical Ping An Technology Shenzhen Co Ltd
Priority to CN201811133725.2A priority Critical patent/CN109145107B/en
Publication of CN109145107A publication Critical patent/CN109145107A/en
Application granted granted Critical
Publication of CN109145107B publication Critical patent/CN109145107B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Machine Translation (AREA)

Abstract

The present invention provides a kind of subject distillation method, apparatus, medium and equipment based on convolutional neural networks, wherein this method comprises: obtaining the term vector matrix of text to be extracted relevant to network public-opinion;According to term vector matrix construction initial characteristics matrix, initial characteristics matrix is extracted into the input of model as topic and is input to the region unit of the first cis-position, and determines the output of region unit;The output of input every other hidden layer in region unit of each hidden layer of region unit;Input by the output of current region block as next region unit continues the output for determining all areas block;According to the revised topic weights of the output of all areas block, the keyword of text to be extracted is extracted according to revised topic weights.The network structure that this method uses can make the transmitting of network characterization and gradient more efficient, avoid gradient disappearance problem caused by loss function information is successively transmitted, and ensure that can be avoided gradient disappearance problem while expanding network depth.

Description

Subject distillation method, apparatus, medium and equipment based on convolutional neural networks
Technical field
The present invention relates to subject distillation technical field, in particular to a kind of subject distillation side based on convolutional neural networks Method, device, medium and equipment.
Background technique
With the development of development of Mobile Internet technology, the network information is in the growth of explosion type, has been full of in network a large amount of useful Or useless text;For example, a kind of form of expression of the network public-opinion as public opinion, the public is based on internet to instantly popular Social concern deliver different view or network public opinion.Since network text information content is huge, rapidly extracting network is needed to believe The main information of breath extracts the theme or abstract of information, to facilitate user quickly to position oneself interested content.
Current subject distillation model is generally basede on based on bag of words and Recognition with Recurrent Neural Network model, and bag of words do not have There is the positional factor for considering word, and text feature is 0 rank statistics;And the computational efficiency of Recognition with Recurrent Neural Network is lower, parameter compared with Do not allow parameter easy adjusting more, and as to will lead to gradient smaller and smaller for continuous iteration, that is, gradient disappearance problem occurs.If using tradition Convolutional neural networks model (CNN) merely increases network layer, also results in the problem of gradient is disappeared with accuracy rate decline, the party Formula can not improve the effect of subject distillation.
Summary of the invention
The present invention provides a kind of subject distillation method, apparatus, medium and equipment based on convolutional neural networks, to solve It is existing to there are problems that the defect of gradient disappearance using the subject distillation model of convolutional neural networks.
A kind of subject distillation method based on convolutional neural networks provided by the invention, comprising:
Text to be extracted relevant to network public-opinion is obtained, each word in the text to be extracted is successively converted into word Vector, and determine the term vector matrix of the text to be extracted, one is distributed for indicating based on corresponding words for each term vector Inscribe the topic weights of keyword possibility size;
According to the term vector matrix construction initial characteristics matrix, using the initial characteristics matrix as the theme after training The input of model is extracted, the subject distillation model includes sequentially connected region unit and connect with all areas block output end Full articulamentum, the output of the full articulamentum are the output of the subject distillation model;
Input by the input of the subject distillation model as the region unit of the first cis-position, and determine the region unit Output;The region unit includes multiple hidden layers, and the input of each hidden layer is every other hidden in the region unit Output containing layer;
Input by the output of current region block as next region unit, continues the output for determining next region unit, Until determining the output of all areas block, and the output of all areas block is reached into full articulamentum;The full articulamentum is according to institute There is the output of region unit to generate the revised topic weights of each term vector of the text to be extracted;
The keyword of the text to be extracted is extracted according to the revised topic weights of term vector.
In one possible implementation, described using the initial characteristics matrix as the subject distillation mould after training Before the input of type, this method further include:
Initial model is constructed, the initial model includes sequentially connected region unit and connect with all areas block output end Full articulamentum, the output of full articulamentum is the output of the initial model;
It obtains term vector and presets matrix, the term vector presets the corresponding topic weights of each term vector in matrix;It is logical It crosses using the default matrix of the term vector as the input of the initial model, by corresponding revised topic weights as described in The output of initial model is trained the initial model, determines the model parameter of the initial model, will determine model ginseng Several initial models is as subject distillation model.
In one possible implementation, the term vector matrix of the determination text to be extracted includes:
Each word in the text to be sorted is successively converted into term vector, and by a sentence of the text to be extracted All term vector sequential concatenations form corresponding sentence sequence in son;
Zero padding processing in tail portion carried out to all sentence sequence, treated that sentence sequence has is identical for tail portion zero padding Sequence length;
Successively using all tail portion zero paddings treated sentence sequence as a row or column of matrix, generate described wait mention Take the term vector matrix of text.
In one possible implementation, described to distribute one for indicating that corresponding words are the theme pass for each term vector The topic weights of keyword possibility size include:
One is distributed for indicating that corresponding words are the theme the identical theme of keyword possibility size for each term vector Weight;Or
It is that term vector distributes one for indicating that corresponding words are the theme keyword possibility size according to the word frequency of term vector Topic weights, the word frequency of the topic weights of the term vector and the term vector is positive correlation.
In one possible implementation, the output of the determination region unit includes:
Preset the processing sequence of all hidden layers in the region unit;
The output of the hidden layer of the first cis-position in the processing sequence is determined according to the input of the region unit, later according to The processing sequence, according to the input of the region unit and have determined the hidden layer before output output successively determine it is every The output of a hidden layer;
According to the processing sequence, successively according to the output of every other hidden layer in the region unit to current hidden layer Output be updated;After updating preset times, by the updated output of hidden layer of cis-position last in the processing sequence Output as the region unit.
In one possible implementation, described to own according to the processing sequence, successively according in the region unit The output of other hidden layers is updated the output of current hidden layer
The output of each hidden layer is successively determined according to the processing sequence, according to more new formula;The more new formula are as follows:
Wherein,Indicate that i-th of hidden layer is in the updated output of kth time in the processing sequence, g () indicates to activate Function, * indicate convolution algorithm;WmiIndicate the weight between m-th of hidden layer and i-th of hidden layer,It indicates to imply for m-th Layer is in the updated output of kth time, WniIndicate the weight between n-th of hidden layer and i-th of hidden layer,It indicates n-th Hidden layer is in kth -1 time updated output;And as k=1, -1 update of kth expression does not update.
In one possible implementation, described to extract the text to be extracted according to the revised topic weights of term vector This keyword includes:
Revised topic weights are greater than and preset term vector corresponding to weight as target term vector, by the target Keyword of the corresponding word of term vector as the text to be extracted;Or
Revised topic weights are ranked up, by word corresponding to the topic weights of preset quantity maximum after sequence Vector is as target term vector, using the corresponding word of the target term vector as the keyword of the text to be extracted.
Based on same inventive concept, the present invention also provides a kind of subject distillation device based on convolutional neural networks, packet It includes:
Module is obtained, it, successively will be in the text to be extracted for obtaining text to be extracted relevant to network public-opinion Each word is converted to term vector, and determines the term vector matrix of the text to be extracted, distributes one for each term vector and is used for Indicate that corresponding words are the theme the topic weights of keyword possibility size;
Determining module is inputted, is used for according to the term vector matrix construction initial characteristics matrix, by the initial characteristics square Battle array as training after subject distillation model input, the subject distillation model include sequentially connected region unit and with own The full articulamentum of region unit output end connection, the output of the full articulamentum are the output of the subject distillation model;
Export determining module, for by the subject distillation model input as the first cis-position region unit input, And determine the output of the region unit;The region unit includes multiple hidden layers, and the input of each hidden layer is from described The output of every other hidden layer in region unit;
Global treatment module, the input by the output of current region block as next region unit, continues to determine next The output of all areas block until determining the output of all areas block, and is reached full articulamentum by the output of region unit;It is described complete Articulamentum generates the revised topic weights of each term vector of the text to be extracted according to the output of all areas block;
Subject distillation module, for extracting the key of the text to be extracted according to the revised topic weights of term vector Word.
Based on same inventive concept, the present invention also provides a kind of computer storage medium, the computer storage medium Computer executable instructions are stored with, the computer executable instructions are for executing method described in above-mentioned any one.
Based on same inventive concept, the present invention also provides a kind of electronic equipment, comprising:
At least one processor;And
The memory being connect at least one described processor communication;Wherein,
The memory is stored with the instruction that can be executed by least one described processor, and described instruction is by described at least one A processor executes, so that at least one described processor is able to carry out method described in above-mentioned any one.
A kind of subject distillation method, apparatus, medium and equipment based on convolutional neural networks provided in an embodiment of the present invention, Term vector based on text to be extracted can construct two-dimensional term vector matrix, later using have sequentially connected region unit and with institute The subject distillation model for the full articulamentum for having region unit to be all connected with determines the topic weights of word, and then extracts corresponding key Word.Utilize multiple hidden layer compositing area blocks, it is possible to reduce the quantity of each hidden layer output characteristic pattern, to reduce network ginseng Several quantity;The network structure of the subject distillation model can make the transmitting of network characterization and gradient more efficient, network Just it is more easier to train;And gradient disappearance problem caused by loss function information is successively transmitted is avoided, it ensure that expansion network It can be avoided gradient disappearance problem while depth, improve subject distillation model training efficiency.With the term vector of each sentence The a line of sequence as term vector matrix, it is ensured that subsequent multistage convolutional calculation.Meanwhile the defeated of hidden layer is determined stage by stage Out, and in second stage by input of the output as current hidden layer of other hidden layers, and based on it is all be updated it is hidden The output for exporting to determine the hidden layer of last cis-position containing layer, and the output as the region unit, can utmostly guarantee The network characteristic of the output of region unit;And the characteristic pattern dimension of region unit will not superlinearity increase, can reduce parameter amount and Calculation amount.
Other features and advantages of the present invention will be illustrated in the following description, also, partly becomes from specification It obtains it is clear that understand through the implementation of the invention.The objectives and other advantages of the invention can be by written explanation Specifically noted structure is achieved and obtained in book, claims and attached drawing.
Below by drawings and examples, technical scheme of the present invention will be described in further detail.
Detailed description of the invention
Attached drawing is used to provide further understanding of the present invention, and constitutes part of specification, with reality of the invention It applies example to be used to explain the present invention together, not be construed as limiting the invention.In the accompanying drawings:
Fig. 1 is the subject distillation method flow diagram based on convolutional neural networks in the embodiment of the present invention;
Fig. 2 is the structural schematic diagram of convolutional neural networks in the embodiment of the present invention;
Fig. 3 is the method flow diagram that term vector matrix is generated in the embodiment of the present invention;
Fig. 4 is the flow diagram that region unit output is determined in the embodiment of the present invention;
Fig. 5 is the subject distillation structure drawing of device based on convolutional neural networks in the embodiment of the present invention;
Fig. 6 is the structure chart of the subject distillation electronic equipment based on convolutional neural networks in the embodiment of the present invention.
Specific embodiment
Hereinafter, preferred embodiments of the present invention will be described with reference to the accompanying drawings, it should be understood that preferred reality described herein Apply example only for the purpose of illustrating and explaining the present invention and is not intended to limit the present invention.
A kind of subject distillation method based on convolutional neural networks provided in an embodiment of the present invention, shown in Figure 1, packet It includes:
Step 101: obtaining text to be extracted relevant to network public-opinion, successively convert each word in text to be extracted It for term vector, and determines the term vector matrix of text to be extracted, distributes one for indicating based on corresponding words for each term vector Inscribe the topic weights of keyword possibility size.
In the embodiment of the present invention, text to be extracted is the text for needing to extract keyword or theme relevant to network public-opinion This, the text may include one or more sentences, and each sentence includes one or more words;Wherein, the corresponding word of each word Vector, and then corresponding term vector matrix can be generated.Wherein, the corresponding topic weights of each term vector, the topic weights For indicating that corresponding words are the theme keyword possibility size, topic weights are bigger, which is more likely to be keyword.Step The topic weights distributed in 101 are initial topic weights, specifically, can distribute one for each term vector for indicating phase Word is answered to be the theme the identical topic weights of keyword possibility size.For example, the topic weights of each term vector are 0.01; Alternatively, the topic weights of each term vector are determined according to the number of term vector, for example, the topic weights of each term vector are 1/N, N are the number of term vector.
Alternatively, being that term vector distributes one for indicating corresponding words according to the word frequency (TF, TermFrequency) of term vector The word frequency of the topic weights for the keyword possibility size that is the theme, the topic weights of the term vector and the term vector is positive The word frequency of pass relationship, i.e. term vector is bigger, and the topic weights primarily determined in step 101 are bigger.For example, text to be extracted is total N number of term vector, some term vector A share a, then the topic weights of term vector A can be a/N.
Step 102: according to term vector matrix construction initial characteristics matrix, using initial characteristics matrix as the theme after training Extract the input of model, subject distillation model includes sequentially connected region unit and what is connect with all areas block output end complete connect Layer is connect, the output of full articulamentum, which is the theme, extracts the output of model.
In the embodiment of the present invention, term vector matrix is equivalent to a gray level image, can benefit when constructing initial characteristics matrix It is constructed with preset multiple convolution kernels;For example, term vector matrix, which is carried out inner product operation with convolution kernel, may thereby determine that Characteristic pattern (feature map), the corresponding initial characteristics matrix of characteristic pattern, the dimension of the convolution kernel can for 5 × 5, Or 6 × 1 etc., the present embodiment does not limit this.In convolutional neural networks, multiple convolution kernels can be set, therefore can construct Multiple initial characteristics matrixes later mention all initial characteristics matrixes as the theme after the training based on convolutional neural networks The input of modulus type.
In the embodiment of the present invention, the primary structure of subject distillation model is shown in Figure 2, which includes more A sequentially connected region unit and the full articulamentum being connected with all areas block.In Fig. 2, with 3 region units (B1, B2, B3 illustrate for), include four hidden layers (h1, h2, h3, h4) in each region unit, the quantity of region unit and hidden layer can root Depending on concrete condition, the present embodiment is not limited this.Multiple regions block is sequentially connected, i.e., the output of a upper region unit can Using the input as next region unit, the output of all areas block is connected to full articulamentum, is mentioned by full articulamentum output theme The output result Output of modulus type.
Wherein, subject distillation model is trained in advance, to determine the model parameter for being appropriate for subject distillation.Specifically , the process of training subject distillation model includes: building initial model, initial model include sequentially connected region unit and with institute The full articulamentum for having region unit output end to connect, the output of full articulamentum are the output of initial model.Building initial model it Afterwards, it obtains term vector and presets matrix, term vector presets the corresponding topic weights of each term vector in matrix;By by term vector Default matrix as the input of initial model, using revised topic weights are as the output of initial model accordingly, to initial Model is trained, and determines the model parameter of initial model, and the initial model of model parameter will be determined as subject distillation model.
In the embodiment of the present invention, which is the subject distillation model before training, initial model and subject distillation mould The network structure of type is identical, and only model parameter may be different, and suitable model parameter is determined by training process.Specifically , term vector is preset into matrix and corresponding revised topic weights as training sample, i.e., it is first that term vector, which presets matrix, The output of the input of beginning model, revised topic weights as initial model adjusts subject distillation model by training Model parameter, the model parameter are specifically as follows network weight, than weight W described as followsmi、WniDeng.Wherein, the term vector is pre- If matrix can be to be determined based on a preset training text according to step 101.
Step 103: the input by the input of subject distillation model as the region unit of the first cis-position, and determine region unit Output;Region unit include multiple hidden layers, and the input of each hidden layer in region unit every other hidden layer it is defeated Out.
In the embodiment of the present invention, the region unit of the first cis-position is first in all areas block being sequentially connected, such as Region unit B1 in Fig. 2.Traditional convolutional neural networks be by input of the output as next hidden layer of a upper hidden layer, i.e., with Hidden layer be unit carry out convolution operation, hidden layer (or convolutional layer) each in this way output feature map quantity compared with Greatly, generally cause convolutional neural networks parameter more with the width (weight) of several hundred or thousands.And comprising more in region unit A hidden layer (being 4 hidden layers in Fig. 2), in this way, the quantity of the output feature map of each hidden layer can be set it is smaller (less than 100), so as to reduce the quantity of network parameter.
Meanwhile in general convolutional neural networks, the input of hidden layer is only related to hidden layer before, exemplary with Fig. 2 Illustrate, in general convolutional neural networks, the input of hidden layer h2 is only related to the output of h1;And in embodiments of the present invention, The output of the input of each hidden layer every other hidden layer in region unit.Such as in Fig. 2, the input of hidden layer h2 with The output of hidden layer h1, h3 and h4 are related.By the output of hidden layer every other in region unit as some hidden layer Input, the result for planting connection type ensures that gradient can be directly accessed from loss function in each hidden layer, can make The transmitting of network characterization and gradient is more efficient, and network is also just more easier to train, it can the very deep network of training.Its In, the output of region unit is the output of some hidden layer;Optionally, the output of region unit is the implicit of last cis-position in hidden layer The output of layer.
Step 104: the input by the output of current region block as next region unit continues to determine next region unit Output and the output of all areas block is reached into full articulamentum until determining the output of all areas block;Full articulamentum according to The output of all areas block generates the revised topic weights of each term vector of text to be extracted.
In the embodiment of the present invention, the input of the region unit of the first cis-position is initial characteristics matrix, the later input of region unit For the output of a upper region unit, can determine according to mode identical in step 103 after the input for determining region unit should The output of region unit.By taking Fig. 2 as an example, the input of region unit B1 is initial characteristics matrix (i.e. Input), and the input of region unit B2 is The output of region unit B1, the input of region unit B3 are the output of region unit B2.Meanwhile each region unit is final with network structure Full articulamentum (Fully Connected layer, FC layer) so that each region unit can directly access loss function letter Breath, and each hidden layer is also capable of the loss function information of direct access region block in each region unit, therefore compared to tradition Only the last one hidden layer is connected with full articulamentum, and it is layer-by-layer that network structure provided in this embodiment avoids loss function information Gradient disappearance problem caused by transmitting, ensure that can be avoided gradient disappearance problem while expanding network depth, improve master Topic extracts model training efficiency, and can train very deep neural network.
Step 105: the keyword of text to be extracted is extracted according to the revised topic weights of term vector.
In the embodiment of the present invention, Text Feature Extraction model is used to correct the topic weights of term vector, each revised master of word Topic weight is mapped in the range of 0~1 to indicate that each word is the probability of keyword.Full articulamentum specifically can be used Sigmoid function or softmax function are realized.Wherein, softmax function is a kind of common more extraction regression models.Sentence It is a two-dimensional problems that whether disconnected target word, which is keyword, and corresponding softmaxt has two dimension, and one-dimensional representation is keyword Probability, two-dimensional representation are not the probability of keyword.
After determining revised topic weights, revised topic weights can be greater than to word corresponding to default weight Vector is as target term vector, using the corresponding word of target term vector as the keyword of text to be extracted.Specifically, by default power Recast is a threshold value, if the topic weights of some word are greater than the threshold value, illustrates that the word has a possibility that sufficiently large as pass Keyword, at this time can be using the word as the keyword of text to be extracted.
Alternatively, revised topic weights are ranked up after determining revised topic weights, it will be maximum after sequence Preset quantity topic weights corresponding to term vector as target term vector, using the corresponding word of target term vector as wait mention Take the keyword of text.For example, this preset quantity is q, then there is the preceding maximum word of q topic weights a possibility that sufficiently large to make It, then can be using the q word as the keyword of text to be extracted for keyword.
A kind of subject distillation method based on convolutional neural networks provided in an embodiment of the present invention, based on text to be extracted Term vector can construct two-dimensional term vector matrix, later using having sequentially connected region unit and be all connected with all areas block The subject distillation model of full articulamentum determines the topic weights of word, and then extracts corresponding keyword.Utilize multiple hidden layers Compositing area block, it is possible to reduce the quantity of each hidden layer output characteristic pattern, to reduce the quantity of network parameter;The theme mentions The network structure of modulus type can make the transmitting of network characterization and gradient more efficient, and network is also just more easier to train;And Gradient disappearance problem caused by loss function information is successively transmitted is avoided, ensure that can be avoided while expanding network depth Gradient disappearance problem improves subject distillation model training efficiency.
The subject distillation method based on convolutional neural networks that another embodiment of the present invention provides a kind of, this method includes above-mentioned Step 101-105 in embodiment, realization principle and technical effect are referring to the corresponding embodiment of Fig. 1.Meanwhile referring to Fig. 3 Shown, in the embodiment of the present invention, step 101 " obtaining text to be extracted relevant to network public-opinion " includes step 1011-1012:
Step 1011: web page text information relevant to network public-opinion, web page text packet are obtained based on web crawlers One or several sentences are included, each sentence includes one or more words.
Step 1012: denoising duplicate removal processing being carried out to web page text information, and to the web page text after denoising duplicate removal processing Information carries out word segmentation processing, using the web page text information after word segmentation processing as text to be extracted.
In the embodiment of the present invention, specifically webpage relevant to network public-opinion, web crawlers skill can be obtained based on web crawlers Art is the technology of comparative maturity, is not repeated herein.Meanwhile after getting web page text information, to web page content information into Row denoising (removing unrelated advertisement etc.) and duplicate removal processing (remove the identical web page contents letter obtained in different url Breath), to reduce the treating capacity of subsequent redundancy.Optionally, weight coefficient, the weight can be arranged to each single item web page text information The initial value of coefficient is 1, and after removing a duplicate web page content information, the weight coefficient of the web page content information adds 1, I.e. weight coefficient is bigger, illustrates that content relevant to the web page text information is more on network, the web page text information it is important Property is higher.It is carrying out word segmentation processing again to web page content information later, stop words is gone to handle, and then can obtain and network public-opinion Relevant phrase set, and using the phrase set as text to be extracted.
Optionally, shown in Figure 3, step 101 " the term vector matrix for determining text to be extracted " includes step 1013- 1015:
Step 1013: each word in text to be extracted being successively converted into term vector, and by one of text to be extracted All term vector sequential concatenations form corresponding sentence sequence in sentence.
In the embodiment of the present invention, each word in text to be extracted can be converted to word2vec term vector, each sentence It is composed of one or more letters, i.e., each corresponding one or more term vectors of sentence are according to sequence of the term vector in sentence Can sequential concatenation formed sentence sequence, the sentence sequence be one-dimension array.
Step 1014: zero padding processing in tail portion being carried out to all sentence sequence, tail portion zero padding treated sentence sequence tool There is identical sequence length.
In the embodiment of the present invention, the full-length of all sentence sequences can be preset, all sentence sequences can also generated Wherein longest sentence sequence is determined afterwards, using the length of the longest sentence sequence as full-length;Determining full-length Afterwards, the sentence sequence inadequate to length carries out tail portion zero padding processing, i.e. last position of subordinate clause subsequence starts to carry out zero padding, directly To reaching full-length.
Step 1015: raw successively using all tail portion zero paddings treated sentence sequence as a row or column of matrix At the term vector matrix of text to be extracted.
Due to zero padding treated sentence sequence sequence length having the same, all sentence sequences have been enumerated at this time A matrix, i.e. term vector matrix can be formed.Under normal circumstances, using sentence sequence as a line of matrix, i.e. term vector The set of the term vector of the corresponding sentence of a line of matrix;At this point, m is text to be extracted for the term vector matrix of m × n In sentence quantity, n is full-length.Optionally, can also preset the size of m and n, i.e. the size of term vector matrix be it is determining, As unit of sentence successively line by line or rows of filling term vector set of matrices later, in term vector matrix not There are the elements of sentence sequence to be set as 0.In the embodiment of the present invention, using the term vector sequence of each sentence as term vector matrix A line, it is ensured that subsequent multistage convolutional calculation.
On the basis of the above embodiments, the process of step 103 " output for determining region unit " is true by two stages Determine the output of region unit, which includes step A1-A3:
Step A1: the processing sequence of all hidden layers in predeterminable area block.
In the embodiment of the present invention, for a region unit, although the input of each hidden layer is owned in region unit The output of other hidden layers, but in actual process, there are processing sequences between hidden layer, implicit to one Layer is disposed after (or temporarily finishing) and handles again next hidden layer.As shown in Fig. 2, the processing of four hidden layers is suitable Sequence can be h1 → h2 → h3 → h4.
Step A2: according to region unit input determine processing sequence in the first cis-position hidden layer output, later according to Processing sequence according to the input of region unit and has determined that the output of the hidden layer before output successively determines each hidden layer Output.
It in the embodiment of the present invention, determines that the output of region unit mainly includes two stages, in the first stage, generates each hidden Characteristic pattern (feature map) containing layer;In second stage, the characteristic pattern of each hidden layer is updated or is adjusted.Specifically , a region unit is shown in Figure 4 in the treatment process in two stages in Fig. 2, and in Fig. 4, dotted portion indicates the first stage, Bold portion indicates second stage;It should be noted that in Fig. 4 for convenience description, property illustrates 8 hidden layers stage by stage, But its network structure is essentially shown in Fig. 2.
In the first stage, the input by the input of region unit as the hidden layer of the first cis-position, and then can determine that this is hidden Output containing layer.With shown in Fig. 4, the hidden layer of the first cis-position is h1, input is the input of the region unit, i.e. X in Fig. 40; When the region unit is region unit (B1 in such as Fig. 2) of the first cis-position, then X0Indicate initial characteristics matrix;When the region unit is When other region units (B2, B3 in such as Fig. 2), then X0For the output of a upper region unit.In the input X for determining hidden layer h10 Afterwards, that is, it can determine the output X of hidden layer h11;Wherein, it in convolutional neural networks, is determined according to the input of hidden layer corresponding defeated It is routine techniques out, the present embodiment is not detailed this.
After determining the output of hidden layer of the first cis-position, the defeated of other hidden layers can be successively determined according to processing sequence Out.Wherein, for second hidden layer h2, the input of region unit is X0, it has been determined that the hidden layer before output only has h1, The input of i.e. second hidden layer h2 includes X0And X1, corresponding weight is distributed for each input, and then can determine the output of h2 For X2;Similarly, the input of third hidden layer h3 includes X0、X1And X2, the output of h3 is X3;The input packet of 4th hidden layer h4 Include X0、X1、X2And X3, the output of h4 is X4.Wherein, X1、X2、X3And X4It is the output of corresponding hidden layer in the first stage, not It is final output.
Step A3: according to processing sequence, successively according to the output of hidden layer every other in region unit to current hidden layer Output be updated;After updating preset times, using the updated output of hidden layer of cis-position last in processing sequence as The output of region unit.
In the embodiment of the present invention, in second stage, successively the output of hidden layer is updated still according to processing sequence; Due to having existed an output in all hidden layers of second stage, (output can be output in the first stage, can also be with For in the last round of updated output of second stage), it can be updated at this time according to the output of all hidden layers.Specifically , the output of each hidden layer is successively determined according to processing sequence, according to more new formula;More new formula are as follows:
Wherein,I-th of hidden layer indicates activation primitive in the updated output of kth time, g () in expression processing sequence, Generally nonlinear activation function, * indicate convolution algorithm;WmiIndicate the weight between m-th of hidden layer and i-th of hidden layer,Indicate m-th of hidden layer in the updated output of kth time, WniIndicate the power between n-th of hidden layer and i-th of hidden layer Weight,Indicate n-th of hidden layer in kth -1 time updated output;And as k=1, -1 update of kth expression does not update. In formula, the maximum value of m and n are the number of hidden layer.
Shown in Figure 4, the first time renewal process after second stage expression in the first stage in Fig. 4 updates public K=1 in formula.According to processing sequence, the input of first hidden layer h1 is the output of hidden layer h2, h3, h4 at this time, i.e. X2、X3With X4, at this time according to X2、X3And X4It can determine the output of h1 after the first round updatesAs k > 1, the input of h1 be hidden layer h2, The output of h3, h4 at this time, only the output of hidden layer h2, h3, h4 at this time is in k-1 updated output.It updates followed by When the output of h2, the output current for other hidden layers of the input of h2 at this time, i.e. the output of hidden layer h1, h3, h4 at this time, at this time The output of h3 and h4 remains as X3And X4, but the output of h1 has been updated toI.e. at this time according to X3、X4WithIt can determine The output of h2 at this timeSimilarly, when updating the output of h3, the input of h3 at this time isAnd X4, updated h3's Output isWhen updating the output of hidden layer h4 of last cis-position, other hidden layers h1, h2, h3 is had been carried out at this time It updates, i.e., according to the updated output of other hidden layersWithDetermine the updated output of h4Work as front-wheel at this time Update operation terminate.If entire renewal process terminates, exported after at this time can updating h4Output as region unit; If also next round is needed to update, k is carried out plus one, repeats the treatment process of second stage in step A3, until k meet it is preset Maximum value.
In the embodiment of the present invention, the output of hidden layer is determined stage by stage, and in second stage by the output of other hidden layers As the input of current hidden layer, and determine based on the output of all hidden layers being updated the hidden layer of last cis-position Output, and the output as the region unit, can utmostly guarantee the network characteristic of the output of region unit;And the spy of region unit Levy figure dimension will not superlinearity increase, parameter amount and calculation amount can be reduced.Meanwhile the input of each hidden layer is region unit Interior all other layer of output can combine other all hidden layers in this way in the block of zoning when the output of a hidden layer Feature map, the result of this connection type ensures that gradient can be directly accessed in each layer from loss function, therefore makes The transmitting for obtaining network characterization is more efficient, and network is also just more easier to train, it can the very deep network of training.
On the basis of the above embodiments, above-mentioned steps 104 are " by the output of current region block as next region unit Input " includes: that the output to current region block carries out pond processing, determines that the characteristic quantity of current region block, characteristic quantity are 1 × 1 The vector of × C, wherein C is the channel number of the output of current region block;Work as proparea according to the adjustment of the characteristic quantity of current region block The output of domain block, the input by the output of current region block adjusted as next region unit.
In the embodiment of the present invention, the output of region unit can be W × H × C matrix, W representing matrix width, H expression square Battle array height, the channel number of C representing matrix;The output is compressed by the output progress pondization processing to region unit, So as to obtain the characteristic quantity of region unit output, while it also can be effectively controlled over-fitting;Later according to current region block Characteristic quantity adjusts the output (such as dot product etc.) of current region block, so that each channel of the output of region unit adjusted is added The characteristic quantity of the upper region unit, so that the characteristic mass for being transferred to next region unit is higher.Wherein, pondization processing specifically may be used It is handled using global poolization;The process of adjustment region block output can also be according to SN-net (Squeeze-and-Excitation Networks) the SN module provided is handled.
On the basis of the above embodiments, step 104 " output of all areas block is reached full articulamentum " includes: by institute There is outputting and inputting for region unit to reach full articulamentum.
In the embodiment of the present invention, outputting and inputting each region unit to be stitched together and carry out global pool, thus often A region unit obtains a corresponding vector, then all areas block global pool result is stitched together so as to execute most Subject distillation process afterwards.Since loss function is determined according to all areas block, then each region block can directly access ladder Information is spent, gradient disappearance problem caused by loss function information is successively transmitted is avoided.
A kind of subject distillation method based on convolutional neural networks provided in an embodiment of the present invention, based on text to be extracted Term vector can construct two-dimensional term vector matrix, later using having sequentially connected region unit and be all connected with all areas block The subject distillation model of full articulamentum determines the topic weights of word, and then extracts corresponding keyword.Utilize multiple hidden layers Compositing area block, it is possible to reduce the quantity of each hidden layer output characteristic pattern, to reduce the quantity of network parameter;The theme mentions The network structure of modulus type can make the transmitting of network characterization and gradient more efficient, and network is also just more easier to train;And Gradient disappearance problem caused by loss function information is successively transmitted is avoided, ensure that can be avoided while expanding network depth Gradient disappearance problem improves subject distillation model training efficiency.Using the term vector sequence of each sentence as term vector matrix A line, it is ensured that it is subsequent multistage convolutional calculation.Meanwhile stage by stage determine hidden layer output, and second stage by its Input of the output of its hidden layer as current hidden layer, and determined finally based on the output of all hidden layers being updated The output of the hidden layer of cis-position, and the output as the region unit can utmostly guarantee that the network of the output of region unit is special Property;And the characteristic pattern dimension of region unit will not superlinearity increase, parameter amount and calculation amount can be reduced.
The subject distillation method flow based on convolutional neural networks is described in detail above, and this method can also be by corresponding Device realize, the structure and function of the device is described in detail below.
A kind of subject distillation device based on convolutional neural networks provided in an embodiment of the present invention, shown in Figure 5, packet It includes:
Module 51 is obtained, it, successively will be in the text to be extracted for obtaining text to be extracted relevant to network public-opinion Each word be converted to term vector, and determine the term vector matrix of the text to be extracted, distribute a use for each term vector In indicating that corresponding words are the theme the topic weights of keyword possibility size;
Determining module 52 is inputted, is used for according to the term vector matrix construction initial characteristics matrix, by the initial characteristics Matrix as training after subject distillation model input, the subject distillation model include sequentially connected region unit and with institute The full articulamentum for having region unit output end to connect, the output of the full articulamentum are the output of the subject distillation model;
Export determining module 53, for by the subject distillation model input as the first cis-position region unit it is defeated Enter, and determines the output of the region unit;The region unit includes multiple hidden layers, and the input of each hidden layer is from institute State the output of every other hidden layer in region unit;
Global treatment module 54, the input by the output of current region block as next region unit, continues to determine next The output of all areas block until determining the output of all areas block, and is reached full articulamentum by the output of a region unit;It is described Full articulamentum generates the revised topic weights of each term vector of the text to be extracted according to the output of all areas block;
Subject distillation module 55, for extracting the key of the text to be extracted according to the revised topic weights of term vector Word.
On the basis of the above embodiments, the acquisition module 51 includes:
Text acquiring unit, for obtaining web page text information relevant to network public-opinion, the net based on web crawlers Page text information includes one or several sentences, and each sentence includes one or more words;
Word segmentation processing unit, for carrying out denoising duplicate removal processing to the web page text information, and to denoising duplicate removal processing Web page text information afterwards carries out word segmentation processing, using the web page text information after word segmentation processing as text to be extracted.
On the basis of the above embodiments, the acquisition module 51 includes:
Converting unit, for each word in the text to be extracted to be successively converted to term vector, and by described wait mention All term vector sequential concatenations in a sentence of text are taken to form corresponding sentence sequence;
Zero padding unit, for carrying out tail portion zero padding processing to all sentence sequence, tail portion zero padding treated sentence Subsequence sequence length having the same;
Matrix generation unit, for successively using all tail portion zero paddings treated sentence sequence as a line of matrix or One column, generate the term vector matrix of the text to be extracted.
On the basis of the above embodiments, described using the initial characteristics matrix as the subject distillation model after training Input before, which further includes training module;
The training module is used for: building initial model, the initial model include sequentially connected region unit and with institute The full articulamentum for having region unit output end to connect, the output of full articulamentum are the output of the initial model;It is pre- to obtain term vector If matrix, the term vector presets the corresponding topic weights of each term vector in matrix;By the way that the term vector is preset square Battle array as the initial model input, using accordingly revised topic weights as the output of the initial model, to institute It states initial model to be trained, determines the model parameter of the initial model, the initial model of model parameter will be determined as master Topic extracts model.
On the basis of the above embodiments, the acquisition module is that each term vector distributes one for indicating that corresponding words are The topic weights of subject key words possibility size include:
One is distributed for indicating that corresponding words are the theme the identical theme of keyword possibility size for each term vector Weight;Or
It is that term vector distributes one for indicating that corresponding words are the theme keyword possibility size according to the word frequency of term vector Topic weights, the word frequency of the topic weights of the term vector and the term vector is positive correlation.
On the basis of the above embodiments, the output determining module 53 includes:
Sequencing unit, for presetting the processing sequence of all hidden layers in the region unit;
Determination unit is exported, determines the implicit of the first cis-position in the processing sequence for the input according to the region unit The output of layer, later according to the processing sequence, implicit according to the input of the region unit and before having determined output The output of layer successively determines the output of each hidden layer;
Updating unit is exported, for according to the processing sequence, successively according to every other hidden layer in the region unit Output the output of current hidden layer is updated;After updating preset times, by cis-position last in the processing sequence Output of the updated output of hidden layer as the region unit.
On the basis of the above embodiments, the output updating unit is used for:
The output of each hidden layer is successively determined according to the processing sequence, according to more new formula;The more new formula are as follows:
Wherein,Indicate that i-th of hidden layer is in the updated output of kth time in the processing sequence, g () indicates to activate Function, * indicate convolution algorithm;WmiIndicate the weight between m-th of hidden layer and i-th of hidden layer,It indicates to imply for m-th Layer is in the updated output of kth time, WniIndicate the weight between n-th of hidden layer and i-th of hidden layer,It indicates n-th Hidden layer is in kth -1 time updated output;And as k=1, -1 update of kth expression does not update.
On the basis of the above embodiments, the Global treatment module 54 includes:
Pond unit carries out pond processing for the output to current region block, determines the characteristic quantity of current region block, institute State the vector that characteristic quantity is 1 × 1 × C, wherein C is the channel number of the output of current region block;
Adjustment unit is worked as adjusting the output of current region block according to the characteristic quantity of current region block by adjusted Input of the output of forefoot area block as next region unit.
On the basis of the above embodiments, the subject distillation module is used for:
Revised topic weights are greater than and preset term vector corresponding to weight as target term vector, by the target Keyword of the corresponding word of term vector as the text to be extracted;Or
Revised topic weights are ranked up, by word corresponding to the topic weights of preset quantity maximum after sequence Vector is as target term vector, using the corresponding word of the target term vector as the keyword of the text to be extracted.
A kind of subject distillation device based on convolutional neural networks provided in an embodiment of the present invention, based on text to be extracted Term vector can construct two-dimensional term vector matrix, later using having sequentially connected region unit and be all connected with all areas block The subject distillation model of full articulamentum determines the topic weights of word, and then extracts corresponding keyword.Utilize multiple hidden layers Compositing area block, it is possible to reduce the quantity of each hidden layer output characteristic pattern, to reduce the quantity of network parameter;The theme mentions The network structure of modulus type can make the transmitting of network characterization and gradient more efficient, and network is also just more easier to train;And Gradient disappearance problem caused by loss function information is successively transmitted is avoided, ensure that can be avoided while expanding network depth Gradient disappearance problem improves subject distillation model training efficiency.Using the term vector sequence of each sentence as term vector matrix A line, it is ensured that it is subsequent multistage convolutional calculation.Meanwhile stage by stage determine hidden layer output, and second stage by its Input of the output of its hidden layer as current hidden layer, and determined finally based on the output of all hidden layers being updated The output of the hidden layer of cis-position, and the output as the region unit can utmostly guarantee that the network of the output of region unit is special Property;And the characteristic pattern dimension of region unit will not superlinearity increase, parameter amount and calculation amount can be reduced.
The embodiment of the present application also provides a kind of computer storage medium, the computer storage medium is stored with computer Executable instruction, it includes the programs for executing the above-mentioned subject distillation method based on convolutional neural networks, which can Execute instruction the method that can be performed in above-mentioned any means embodiment.
Wherein, the computer storage medium can be any usable medium that computer can access or data storage is set It is standby, including but not limited to magnetic storage (such as floppy disk, hard disk, tape, magneto-optic disk (MO) etc.), optical memory (such as CD, DVD, BD, HVD etc.) and semiconductor memory (such as ROM, EPROM, EEPROM, nonvolatile memory (NAND FLASH), solid state hard disk (SSD)) etc..
Fig. 6 shows the structural block diagram of a kind of electronic equipment of another embodiment of the invention.The electronic equipment 1100 can be the host server for having computing capability, personal computer PC or portable portable computer or end End etc..The specific embodiment of the invention does not limit the specific implementation of electronic equipment.
The electronic equipment 1100 includes at least one processor (processor) 1110, communication interface (Communications Interface) 1120, memory (memory array) 1130 and bus 1140.Wherein, processor 1110, communication interface 1120 and memory 1130 complete mutual communication by bus 1140.
Communication interface 1120 with network element for communicating, and wherein network element includes such as Virtual Machine Manager center, shared storage.
Processor 1110 is for executing program.Processor 1110 may be a central processor CPU or dedicated collection At circuit ASIC (Application Specific Integrated Circuit), or it is arranged to implement the present invention One or more integrated circuits of embodiment.
Memory 1130 is for executable instruction.Memory 1130 may include high speed RAM memory, it is also possible to also wrap Include nonvolatile memory (non-volatile memory), for example, at least a magnetic disk storage.Memory 1130 can also be with It is memory array.Memory 1130 is also possible to by piecemeal, and described piece can be combined into virtual volume by certain rule.Storage The instruction that device 1130 stores can be executed by processor 1110, so that processor 1110 is able to carry out in above-mentioned any means embodiment Method.
Obviously, various changes and modifications can be made to the invention without departing from essence of the invention by those skilled in the art Mind and range.In this way, if these modifications and changes of the present invention belongs to the range of the claims in the present invention and its equivalent technologies Within, then the present invention is also intended to include these modifications and variations.

Claims (10)

1. a kind of subject distillation method based on convolutional neural networks characterized by comprising
Obtain relevant to network public-opinion text to be extracted, successively by each word in the text to be extracted be converted to word to Amount, and determine the term vector matrix of the text to be extracted, one is distributed for indicating that corresponding words are the theme for each term vector The topic weights of keyword possibility size;
According to the term vector matrix construction initial characteristics matrix, using the initial characteristics matrix as the subject distillation after training The input of model, the subject distillation model include sequentially connected region unit and connect with all areas block output end entirely connect Layer is connect, the output of the full articulamentum is the output of the subject distillation model;
Input by the input of the subject distillation model as the region unit of the first cis-position, and determine the defeated of the region unit Out;The region unit includes multiple hidden layers, and the input of each hidden layer is every other implicit in the region unit The output of layer;
Input by the output of current region block as next region unit, continues the output for determining next region unit, until It determines the output of all areas block, and the output of all areas block is reached into full articulamentum;The full articulamentum is according to all areas The output of domain block generates the revised topic weights of each term vector of the text to be extracted;
The keyword of the text to be extracted is extracted according to the revised topic weights of term vector.
2. the method according to claim 1, wherein it is described using the initial characteristics matrix as training after Before the input of subject distillation model, further includes:
Initial model is constructed, the initial model includes sequentially connected region unit and connect with all areas block output end complete Articulamentum, the output of full articulamentum are the output of the initial model;
It obtains term vector and presets matrix, the term vector presets the corresponding topic weights of each term vector in matrix;Passing through will The term vector presets matrix as the input of the initial model, by corresponding revised topic weights as described initial The output of model is trained the initial model, determines the model parameter of the initial model, will determine model parameter Initial model is as subject distillation model.
3. the method according to claim 1, wherein the term vector matrix packet of the determination text to be extracted It includes:
Each word in the text to be sorted is successively converted into term vector, and will be in a sentence of the text to be extracted All term vector sequential concatenations form corresponding sentence sequence;
Zero padding processing in tail portion carried out to all sentence sequence, tail portion zero padding treated sentence sequence sequence having the same Column length;
Successively using all tail portion zero paddings treated sentence sequence as a row or column of matrix, the text to be extracted is generated This term vector matrix.
4. the method according to claim 1, wherein described distribute one for indicating corresponding for each term vector The be the theme topic weights of keyword possibility size of word include:
One is distributed for indicating that corresponding words are the theme the identical topic weights of keyword possibility size for each term vector; Or
It is that term vector distributes one for indicating that corresponding words are the theme the master of keyword possibility size according to the word frequency of term vector The word frequency of topic weight, the topic weights of the term vector and the term vector is positive correlation.
5. method according to claim 1 to 4, which is characterized in that the output of the determination region unit includes:
Preset the processing sequence of all hidden layers in the region unit;
The output that the hidden layer of the first cis-position in the processing sequence is determined according to the input of the region unit, later according to described Processing sequence, according to the input of the region unit and have determined the hidden layer before output output successively determine it is each hidden Output containing layer;
According to the processing sequence, successively according to the output of every other hidden layer in the region unit to the defeated of current hidden layer It is updated out;After updating preset times, using the updated output of hidden layer of cis-position last in the processing sequence as The output of the region unit.
6. according to the method described in claim 4, it is characterized in that, described according to the processing sequence, successively according to the area The output of every other hidden layer, which is updated the output of current hidden layer, in the block of domain includes:
The output of each hidden layer is successively determined according to the processing sequence, according to more new formula;The more new formula are as follows:
Wherein,Indicate that i-th of hidden layer is in the updated output of kth time in the processing sequence, g () indicates activation letter Number, * indicate convolution algorithm;WmiIndicate the weight between m-th of hidden layer and i-th of hidden layer,Indicate m-th of hidden layer In the updated output of kth time, WniIndicate the weight between n-th of hidden layer and i-th of hidden layer,Indicate n-th it is hidden Containing layer in kth -1 time updated output;And as k=1, -1 update of kth expression does not update.
7. method according to claim 1 to 4, which is characterized in that described according to the revised topic weights of term vector The keyword for extracting the text to be extracted includes:
Using revised topic weights be greater than default weight corresponding to term vector as target term vector, by the target word to Measure keyword of the corresponding word as the text to be extracted;Or
Revised topic weights are ranked up, by term vector corresponding to the topic weights of preset quantity maximum after sequence As target term vector, using the corresponding word of the target term vector as the keyword of the text to be extracted.
8. a kind of subject distillation device based on convolutional neural networks characterized by comprising
Module is obtained, for obtaining text to be extracted relevant to network public-opinion, successively by each of described text to be extracted Word is converted to term vector, and determines the term vector matrix of the text to be extracted, distributes one for indicating for each term vector Corresponding words are the theme the topic weights of keyword possibility size;
Determining module is inputted, for according to the term vector matrix construction initial characteristics matrix, the initial characteristics matrix to be made For the input of the subject distillation model after training, the subject distillation model includes sequentially connected region unit and and all areas The full articulamentum of block output end connection, the output of the full articulamentum are the output of the subject distillation model;
Determining module is exported, for the input by the input of the subject distillation model as the region unit of the first cis-position, and really The output of the fixed region unit;The region unit includes multiple hidden layers, and the input of each hidden layer is from the region The output of every other hidden layer in block;
Global treatment module, the input by the output of current region block as next region unit, continues to determine next region The output of all areas block until determining the output of all areas block, and is reached full articulamentum by the output of block;The full connection Layer generates the revised topic weights of each term vector of the text to be extracted according to the output of all areas block;
Subject distillation module, for extracting the keyword of the text to be extracted according to the revised topic weights of term vector.
9. a kind of computer storage medium, which is characterized in that the computer storage medium is stored with computer executable instructions, The computer executable instructions require method described in 1-7 any one for perform claim.
10. a kind of electronic equipment characterized by comprising
At least one processor;And
The memory being connect at least one described processor communication;Wherein,
The memory is stored with the instruction that can be executed by least one described processor, and described instruction is by described at least one It manages device to execute, so that at least one described processor is able to carry out method described in claim 1-7 any one.
CN201811133725.2A 2018-09-27 2018-09-27 Theme extraction method, device, medium and equipment based on convolutional neural network Active CN109145107B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811133725.2A CN109145107B (en) 2018-09-27 2018-09-27 Theme extraction method, device, medium and equipment based on convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811133725.2A CN109145107B (en) 2018-09-27 2018-09-27 Theme extraction method, device, medium and equipment based on convolutional neural network

Publications (2)

Publication Number Publication Date
CN109145107A true CN109145107A (en) 2019-01-04
CN109145107B CN109145107B (en) 2023-07-25

Family

ID=64813025

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811133725.2A Active CN109145107B (en) 2018-09-27 2018-09-27 Theme extraction method, device, medium and equipment based on convolutional neural network

Country Status (1)

Country Link
CN (1) CN109145107B (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109886390A (en) * 2019-01-10 2019-06-14 平安科技(深圳)有限公司 Convolutional neural networks model optimization method, apparatus, computer equipment and storage medium
CN110991175A (en) * 2019-12-10 2020-04-10 爱驰汽车有限公司 Text generation method, system, device and storage medium under multiple modes
CN111428489A (en) * 2020-03-19 2020-07-17 北京百度网讯科技有限公司 Comment generation method and device, electronic equipment and storage medium
CN111597296A (en) * 2019-02-20 2020-08-28 阿里巴巴集团控股有限公司 Commodity data processing method, device and system

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105678214A (en) * 2015-12-21 2016-06-15 中国石油大学(华东) Vehicle flow statistical method based on convolutional neural network vehicle model recognition in cloud environment
US20170011279A1 (en) * 2015-07-07 2017-01-12 Xerox Corporation Latent embeddings for word images and their semantics
JP2017059205A (en) * 2015-09-17 2017-03-23 パナソニックIpマネジメント株式会社 Subject estimation system, subject estimation method, and program
US20170140253A1 (en) * 2015-11-12 2017-05-18 Xerox Corporation Multi-layer fusion in a convolutional neural network for image classification
CN107665353A (en) * 2017-09-15 2018-02-06 平安科技(深圳)有限公司 Model recognizing method, device, equipment and computer-readable recording medium based on convolutional neural networks
CN108170736A (en) * 2017-12-15 2018-06-15 南瑞集团有限公司 A kind of document based on cycle attention mechanism quickly scans qualitative method
CN108304364A (en) * 2017-02-23 2018-07-20 腾讯科技(深圳)有限公司 keyword extracting method and device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20170011279A1 (en) * 2015-07-07 2017-01-12 Xerox Corporation Latent embeddings for word images and their semantics
JP2017059205A (en) * 2015-09-17 2017-03-23 パナソニックIpマネジメント株式会社 Subject estimation system, subject estimation method, and program
US20170140253A1 (en) * 2015-11-12 2017-05-18 Xerox Corporation Multi-layer fusion in a convolutional neural network for image classification
CN105678214A (en) * 2015-12-21 2016-06-15 中国石油大学(华东) Vehicle flow statistical method based on convolutional neural network vehicle model recognition in cloud environment
CN108304364A (en) * 2017-02-23 2018-07-20 腾讯科技(深圳)有限公司 keyword extracting method and device
CN107665353A (en) * 2017-09-15 2018-02-06 平安科技(深圳)有限公司 Model recognizing method, device, equipment and computer-readable recording medium based on convolutional neural networks
CN108170736A (en) * 2017-12-15 2018-06-15 南瑞集团有限公司 A kind of document based on cycle attention mechanism quickly scans qualitative method

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109886390A (en) * 2019-01-10 2019-06-14 平安科技(深圳)有限公司 Convolutional neural networks model optimization method, apparatus, computer equipment and storage medium
CN109886390B (en) * 2019-01-10 2023-11-24 平安科技(深圳)有限公司 Convolutional neural network model optimization method, device, computer equipment and storage medium
CN111597296A (en) * 2019-02-20 2020-08-28 阿里巴巴集团控股有限公司 Commodity data processing method, device and system
CN110991175A (en) * 2019-12-10 2020-04-10 爱驰汽车有限公司 Text generation method, system, device and storage medium under multiple modes
CN110991175B (en) * 2019-12-10 2024-04-09 爱驰汽车有限公司 Method, system, equipment and storage medium for generating text in multi-mode
CN111428489A (en) * 2020-03-19 2020-07-17 北京百度网讯科技有限公司 Comment generation method and device, electronic equipment and storage medium
CN111428489B (en) * 2020-03-19 2023-08-29 北京百度网讯科技有限公司 Comment generation method and device, electronic equipment and storage medium

Also Published As

Publication number Publication date
CN109145107B (en) 2023-07-25

Similar Documents

Publication Publication Date Title
Kalchbrenner et al. Efficient neural audio synthesis
CN108010514B (en) Voice classification method based on deep neural network
US11645529B2 (en) Sparsifying neural network models
CN109145107A (en) Subject distillation method, apparatus, medium and equipment based on convolutional neural networks
CN107437096B (en) Image classification method based on parameter efficient depth residual error network model
US8131659B2 (en) Field-programmable gate array based accelerator system
CN109543029A (en) File classification method, device, medium and equipment based on convolutional neural networks
CN109871532A (en) Text subject extracting method, device and storage medium
CN110175628A (en) A kind of compression algorithm based on automatic search with the neural networks pruning of knowledge distillation
CN110210618A (en) The compression method that dynamic trimming deep neural network weight and weight are shared
CN105825269B (en) A kind of feature learning method and system based on parallel automatic coding machine
CN111126602A (en) Cyclic neural network model compression method based on convolution kernel similarity pruning
WO2023098544A1 (en) Structured pruning method and apparatus based on local sparsity constraints
CN108647723A (en) A kind of image classification method based on deep learning network
CN112508190A (en) Method, device and equipment for processing structured sparse parameters and storage medium
CN108038539A (en) A kind of integrated length memory Recognition with Recurrent Neural Network and the method for gradient lifting decision tree
CN112529415B (en) Article scoring method based on combined multiple receptive field graph neural network
CN109840585A (en) A kind of operation method and system towards sparse two-dimensional convolution
CN109241298A (en) Semantic data stores dispatching method
CN111259157A (en) Chinese text classification method based on hybrid bidirectional circulation capsule network model
CN106169961A (en) The network parameter processing method and processing device of neutral net based on artificial intelligence
Liu et al. EACP: An effective automatic channel pruning for neural networks
Du et al. Efficient network construction through structural plasticity
CN109583586A (en) A kind of convolution kernel processing method and processing device
CN109033304A (en) Multi-modal retrieval method based on online deep layer topic model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant