CN115392232A - Topic and multi-mode fused emergency emotion analysis method - Google Patents
Topic and multi-mode fused emergency emotion analysis method Download PDFInfo
- Publication number
- CN115392232A CN115392232A CN202211003522.8A CN202211003522A CN115392232A CN 115392232 A CN115392232 A CN 115392232A CN 202211003522 A CN202211003522 A CN 202211003522A CN 115392232 A CN115392232 A CN 115392232A
- Authority
- CN
- China
- Prior art keywords
- text
- emotion
- topic
- picture
- comment
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/237—Lexical tools
- G06F40/242—Dictionaries
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/70—Arrangements for image or video recognition or understanding using pattern recognition or machine learning
- G06V10/764—Arrangements for image or video recognition or understanding using pattern recognition or machine learning using classification, e.g. of video objects
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- General Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- General Engineering & Computer Science (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Evolutionary Computation (AREA)
- Software Systems (AREA)
- Computing Systems (AREA)
- Data Mining & Analysis (AREA)
- Biophysics (AREA)
- Mathematical Physics (AREA)
- Molecular Biology (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Databases & Information Systems (AREA)
- Medical Informatics (AREA)
- Multimedia (AREA)
- Machine Translation (AREA)
Abstract
A topic and multi-mode fused emergent event sentiment analysis method is characterized in that external knowledge is fused in neural topic modeling, a neural topic model obtained through modeling is pre-trained on a large corpus, then fine adjustment is carried out on a target data set, then sentiment distribution of topics is calculated by utilizing a sentiment dictionary, and further the sentiment tendency of each comment in the data set is obtained; based on the neural topic model, the comment to be analyzed is taken as input to obtain the sentiment value M of the comment to be analyzed 1 (ii) a For theAnd performing text and picture correlation analysis to obtain a text correlation coefficient and a picture correlation coefficient to obtain a text emotion value and a picture emotion value, performing weighted average operation according to the text correlation coefficient to obtain a text emotion value M and a picture emotion value M 2 (ii) a Will M 1 And M 2 And carrying out fusion on the model result level to obtain the final emotion value M of the comment. According to the method and the device, sentiment analysis can be more accurately carried out on comment contents in similar websites or platforms such as microblogs.
Description
Technical Field
The invention belongs to the technical field of artificial intelligence, relates to emotion analysis in network events, and particularly relates to an emergent event emotion analysis method integrating themes and multiple modes.
Background
Emotion analysis is a task in the field of natural language processing, also called tendency analysis, opinion extraction, opinion mining, emotion mining, subjective analysis and the like, and is a process of analyzing, processing, inducing and reasoning subjective texts with emotion colors.
With the rapid development of networks, social networks become a main platform for network public opinion propagation, and microblogs and other similar websites or platforms serve as important media for network public opinion propagation, so that users can express opinions anytime and anywhere. Unlike traditional text data, the comment data is redundant and contains texts, pictures, videos and a large amount of special information such as website or platform specific emoticons, and meanwhile, text emotions are closely related to discussion topics, which causes great difficulty in emotion analysis of comment content in similar websites or platforms such as microblogs.
Disclosure of Invention
In order to overcome the defects of the prior art, the invention aims to provide an emergent event emotion analysis method combining themes and multiple modes, so that the comment content in similar websites or platforms such as microblogs can be more accurately analyzed.
In order to achieve the purpose, the invention adopts the technical scheme that:
a topic and multi-mode combined emergent event emotion analysis method comprises the following steps:
step 1, external knowledge is blended in neural topic modeling, and neural topic models obtained through modeling are pre-processed on a large corpusTraining, then carrying out fine adjustment on a target data set, and then calculating topic emotion distribution by using an emotion dictionary to further obtain the emotional tendency of each comment in the data set; the data in the large corpus and the target data set comprise texts and pictures; based on the neural topic model, the comment to be analyzed is taken as input to obtain the sentiment value M of the comment 1 ;
Step 2, performing text and picture correlation analysis on the comment to be analyzed to obtain a picture and text correlation coefficient mu;
step 3, extracting emotion characteristics of the text and the picture by adopting a method of fusing BiLSTM, textCNN and attention mechanism to obtain text emotion valueAnd picture emotion valueAccording to the graphic correlation coefficient mu pairAndcarrying out weighted average operation to obtain the image-text emotion value
Step 4, the emotion value M 1 And an emotion value M 2 And carrying out fusion on the model result level to obtain the final emotion value M of the comment.
In one embodiment, in step 1, the external knowledge is knowledge which is learned about a subject and can be reused in fine tuning on the target data set when the neural subject model is pre-trained; is merged by the neural topic model through pre-training.
In one embodiment, the neural topic model adopts an encoder-decoder architecture, a BoW model is adopted to process texts in a data set to obtain x, and an encoder processes x e R v To input, dataThe topic distribution of the concentrated text is t epsilon R k Wherein, v is vocabulary, k is the subject number of the subject distribution t; the decoder reconstructs the original document; the encoder is a stack consisting of N +1 MLP layers, from bottom to top, the first N layers have the same structure, each layer has four sublayers of Dropout, linear, batchNorm and LeakyReLU, the last layer is a Dropout sublayer and a Linear transform, and then a Softmax layer, and the decoder and the encoder have the same architecture.
In one embodiment, the encoder receives x ∈ R v As input, and deducing the topic distribution t epsilon R thereof k Then, the decoder reconstructs the original document from t, and in the process, the exit probability of each layer of the encoder and the negative slope of the LeakyReLU sublayer are set to obtain the reconstruction loss: l. the rec (x,t)=-E(xlogt);
Wherein t and x have the same size m, and the topic distribution obtained by the neural topic model is adjusted by minimizing the maximum mean difference between dirichlet distribution P, and the formula is as follows:
the overall training goals are: l = l rec (x,t)+r·λ·l MMD (t,t′)
t' is the subject distribution randomly extracted from P, k () is the information diffusion function, i, j take on values from 1 to m;
r is a hyperparameter for balancing l rec And l MMD ,Normalization is performed using a two-norm, b (N + 1) is the deviation term before the Softmax sublayer of the encoder,is the derivation operator.
In one embodiment, the large corpus is a DBPedia data set, and the target data set is a data set of netizen emotion recognition games during epidemic situations of CCIR 2020; the emotion dictionary is an emotion polarity dictionary of Taiwan university; training the neural topic model on the DBPedia data set once to finish pre-training; then, fine tuning is done on the data set of the netizen emotion recognition game during the epidemic situation of CCIR 2020.
In one embodiment, the fine tuning starts from a pre-trained model, parameters are randomly re-initialized at the last layer of an encoder and the first layer of the decoder, emotion values of all themes are obtained according to an emotion dictionary, and then a whole comment emotion value M based on the themes is obtained 1 。
In one embodiment, in step 2, a fusion method of BilSTM and attention mechanism is used for text and picture correlation analysis, and the method comprises the following steps:
firstly, processing a text and an image, converting the text into a text matrix by adopting a Glove method, extracting an image label by utilizing a tool provided by a Vision Platform in a Google Cloud Platform, and expressing the image label into a word matrix the same as the text;
then using two independent BilSTMs to respectively receive the picture label and the text label, and representing the picture and the text as vectors with the same dimensionality through the BilSTM;
finally, feature splicing is carried out on the image vector and the text vector to serve as input of a full connection layer, and finally the image-text correlation coefficient mu is output through a softmax layer.
In one embodiment, in step 3, based on a text classification method of a BilSTM-Attention-textCNN hybrid neural network, a text in a comment to be analyzed is mapped into a vector through a word embedding layer, and a BilSTM network is used to learn an upper expression and a lower expression of a word in the comment to be analyzed, so as to obtain a semantic vector of a deeper current word; establishing an attention model, and calculating the probability weight of each word vector to enable words with larger weights to get more attention, wherein the words with more attention are often key words for classification tasks; connecting the vectors output by the attention mechanism with a pooling layer, performing k-max pooling, and reserving the first k words with larger weights; connecting textCNN network extraction feature output text emotion value
The method comprises the steps of extracting labels from pictures in comments to be analyzed and expressing the labels into matrixes which are the same as texts, extracting picture characteristics by using the BilSTM, establishing an Attention mechanism to select the characteristics, connecting vectors output by the Attention mechanism to a pooling layer, executing k-max pooling, and reserving the first k words with larger weights; connecting CNN network to extract characteristic output picture emotion value
In one embodiment, in step 4, the fused calculation formula of the final emotion value M is as follows:
the specific gravity delta is determined by adjusting parameters in the model training process.
Since the network comments often contain specific information such as texts and pictures, and meanwhile, the comment sentiments are closely related to the discussion subjects. Therefore, compared with the existing emotion analysis method, the emotion analysis method provided by the invention analyzes emotion through the topic and multi-mode fusion analysis method, and is easier to deal with the emotion analysis of the network comment data.
Drawings
FIG. 1 is an architectural diagram of a neural topic model.
Fig. 2 is a diagram showing a graph-text correlation structure.
FIG. 3 is a schematic diagram of text emotion extraction.
Fig. 4 is a schematic diagram of emotional feature extraction.
FIG. 5 is a schematic diagram of a comment emotion polarity acquisition structure.
FIG. 6 is a drawing of a text review in an embodiment of the present invention.
FIG. 7 is a diagram of a second match of text comments in the embodiment of the present invention.
Detailed Description
The embodiments of the present invention will be described in detail below with reference to the drawings and examples.
The invention relates to an emergent event emotion analysis method integrating a theme and multiple modes, which combines the theme and the multiple modes to carry out emotional analysis on an emergent event.
In the invention, the multi-mode refers to text and pictures, and obviously has themes around emergencies.
The implementation of the invention comprises the following steps:
step 1, external knowledge is blended in neural topic modeling, a neural topic model obtained through modeling is pre-trained on a large corpus, then fine tuning is carried out on a target data set, then topic emotion distribution is calculated by using an emotion dictionary, and then the emotional tendency of each comment in the data set is obtained. Based on the neural topic model, the comment to be analyzed is taken as input to obtain the sentiment value M of the comment to be analyzed 1 ;
In the invention, the external knowledge is learned and related to a theme when the neural theme model is pre-trained, and can be repeatedly used when fine tuning is carried out on a target data set, and the external knowledge is merged by the neural theme model through pre-training.
The large corpus used in the method is a DBPedia data set, the target data set is a data set of netizen emotion recognition matches in the epidemic situation period of CCIR2020, the target data set is microblog comment data related to the epidemic situation, and emotion analysis is conducted in the embodiment, namely emotion polarity of each comment is analyzed, namely negative or positive. It should be noted that the data in the large corpus and the target dataset both include text, pictures, expressions, and the like. The invention uses only text and pictures, wherein only text is used in step 1. The emotion dictionary used by the invention is an emotion polarity dictionary of Taiwan university, and the neural topic model is trained on a DBPedia data set once to complete pre-training; the fine-tuning is then done on the target data set.
Referring to FIG. 1, the neural topic model established by the present invention adopts encoder-decoder architecture, firstly adopts BThe oW model processes texts in a data set (DBPedia data set and target data set) to obtain x, and the topic distribution of the texts in the data set is t e R k Where v is the vocabulary, k is the topic number of the topic distribution t, the encoder is a stack of N +1 MLP layers, from bottom to top, the first N layers have the same structure, each layer has four sublayers, namely, dryout sublayer, linear sublayer, batchNorm sublayer and LeakyReLU sublayer, the last layer is a Dropout sublayer and a Linear transform, and then a Softmax layer. The decoder reconstructs the original document, which has the same architecture as the encoder.
Specifically, the encoder receives x ∈ R v As input, and deducing the topic distribution t epsilon R k Then, the decoder reconstructs the original document from t, and in the process, the exit probability of each layer of the encoder and the negative slope of the LeakyReLU sublayer are set to obtain the reconstruction loss: l. the rec (x,t)=-E(xlogt);
Wherein t and x have the same size m, the topic distribution obtained by the neural topic model adjusts the topic by minimizing the maximum mean difference between the Dirichlet distribution P, and the formula is as follows:
the overall training objectives are: l = l rec (x,t)+r·λ·l MMD (t,t′)
t' is the subject distribution randomly extracted from P, k () is the information diffusion function, i, j take on values from 1 to m; r is a hyperparameter for balancing l rec And l MMD ,Normalization is performed using a two-norm, b (N + 1) is the deviation term before the Softmax sublayer of the encoder,is the derivation operator.
The fine-tuning in this step starts with the pre-trained neural topic model, but in the last layer of the encoderAnd a decoder first layer randomly reinitializes the parameters, obtains the emotion value of each theme according to the emotion dictionary, and further obtains the emotion value M of the whole comment based on the theme 1 。
And 2, performing text and picture correlation analysis on a certain comment to be analyzed to obtain a picture and text correlation coefficient mu.
The structure of the teletext relevance is shown in fig. 2, which can be expressed as relevance and irrelevance, if not, only emotion analysis of the text is performed, and if relevant, comprehensive analysis is performed.
In the step, a BilSTM and attention mechanism fusion method is adopted to carry out text and picture correlation analysis, and the method comprises the following steps:
firstly, processing a text and an image, converting the text into a text matrix by adopting a Glove method, extracting an image label by utilizing a tool provided by a Vision Platform in Google Cloud Platform, and expressing the image label into a word matrix the same as the text.
Then two independent BilSTMs are used for respectively receiving the picture labels and the text labels, and the pictures and the texts are represented as vectors with the same dimensionality through the BilSTMs.
Finally, feature splicing is carried out on the picture vectors and the text vectors, the feature splicing is used as input of a full connection layer (namely the full connection layer formed by splicing the text and the picture feature vectors), and finally, the picture and text correlation coefficient mu is output through the softmax layer.
Step 3, extracting emotion characteristics of the text and the picture by adopting a method of fusing BiLSTM, textCNN and attention mechanism to obtain text emotion valueAnd picture emotion valueAccording to the image-text correlation mu pairAndcarrying out weighted average operation to obtain the image-text sentiment value M 2 。
in the step, the Attention Mechanism (Attention Mechanism) is utilized to automatically learn and calculate the contribution of the input data to the output data, so that the extraction of the emotional characteristics is more reasonable.
Specifically, referring to fig. 3, based on the text classification method of the BilSTM-Attention-textCNN hybrid neural network, the text in the comment to be analyzed is mapped into a vector through a word embedding layer, and the BilSTM network is used to learn the upper expression and the lower expression of the word in the comment to be analyzed, so as to obtain a semantic vector of the current word with a deeper level; establishing an attention model, and calculating the probability weight of each word vector to enable words with larger weights to get more attention, wherein the words with more attention are often key words for classification tasks; connecting vectors output by the attention mechanism with a pooling layer, performing k-max pooling, and reserving the former k words with larger weights; connecting textCNN network extraction feature output text emotion value
Referring to fig. 4, in the image classification method based on the BiLSTM-Attention-CNN hybrid neural network, extracting tags from images in a comment to be analyzed and representing the tags into a matrix the same as a text, extracting image features using the BiLSTM, establishing an Attention mechanism to select the features, connecting vectors output by the Attention mechanism to a pooling layer, performing k-max pooling, and reserving the first k words with higher weights; connecting CNN network to extract characteristic output picture emotion value
In the step 4, the step of the method,referring to FIG. 5, the emotion value M 1 And an emotion value M 2 And (3) carrying out fusion on the model result level to obtain the final sentiment value M of the comment, wherein the fusion calculation formula is as follows:
the specific gravity delta is determined by adjusting parameters in the model training process.
In one particular embodiment of the invention, the text comment data is intercepted as follows:
4456427143652010,01, 23/02/3, 21, mimiko sweet heart, no. 1/2, first small fever hoped for this year to be a safe, healthy, and happy 2 shijiazhuang, hebei science and technology university?
Fig. 6 and 7 are diagrams of the text review.
And (4) the text of the target data set is processed by a neural topic model and an emotion dictionary to obtain the emotion value of the topic.
The comment accumulates the sentiment value of the theme according to the theme and then averages, and the sentiment value of the theme contained in the comment is m 1 ,m 2 ,…,m n Then the sentiment value M derived from the subject 1 =(m 1 +m 2 +…+m n )/n。
Inputting the data into the correlation model shown in FIG. 2 to obtain the correlation coefficient μ between the text and the picture in the comment, and inputting the data into the text picture emotion value extraction model shown in FIG. 3 and FIG. 4 to obtain the emotion value between the text and the pictureAndthe emotion value obtained from the image and text is:
will feel the value M 1 And an emotion value M 2 The final sentiment value M of the comment is calculated according to the weight delta, and then the comment is obtainedThe sentiment value of (A) is:
m >0 aggressive; m =0 neutral; m <0 negative.
Claims (9)
1. A topic and multi-mode combined emergent event emotion analysis method is characterized by comprising the following steps:
step 1, external knowledge is blended in neural topic modeling, a neural topic model obtained through modeling is pre-trained on a large corpus, then fine tuning is carried out on a target data set, then topic emotion distribution is calculated by using an emotion dictionary, and then emotion tendency of each comment in the data set is obtained; the data in the large corpus and the target data set comprise texts and pictures; based on the neural topic model, the comment to be analyzed is taken as input to obtain the sentiment value M of the comment to be analyzed 1 ;
Step 2, performing text and picture correlation analysis on the comment to be analyzed to obtain a picture-text correlation coefficient mu;
step 3, extracting emotion characteristics of the text and the picture by adopting a method of fusing BiLSTM, textCNN and attention mechanism to obtain text emotion valueAnd picture emotion valueAccording to the graphic correlation coefficient mu pairAndcarrying out weighted average operation to obtain the image-text emotion value
Step 4, the emotion value M 1 And an emotion value M 2 And (5) carrying out fusion on the model result layer to obtain the final emotion value M of the comment.
2. The method for analyzing emergent events and emotions integrating topics and multimodalities according to claim 1, wherein in the step 1, the external knowledge is knowledge which is learned and related to topics and can be reused when fine-tuning on the target data set when the neural topic model is pre-trained; is merged by the neural topic model through pre-training.
3. The method for analyzing emergent events and emotions integrating themes and multimodal as claimed in claim 1, wherein the neural theme model adopts a coder-decoder architecture, a BoW model is adopted to process texts in a data set to obtain x, and a coder processes x e R v For input, the subject distribution of the text in the data set is t epsilon R k Wherein v is vocabulary, k is the topic number of the topic distribution t; the decoder reconstructs the original document; the encoder is a stack consisting of N +1 MLP layers, from bottom to top, the first N layers have the same structure, each layer has four sublayers, namely, dropout, linear, batchNorm, and LeakyReLU, the last layer is a Dropout sublayer and a Linear transform, and then a Softmax layer, and the decoder and the encoder have the same architecture.
4. The method for topic-fusion and multi-modal analysis of emotion from incident events according to claim 3, wherein said encoder receives x e R v As input, and deducing the topic distribution t epsilon R thereof k Then, the decoder reconstructs the original document from t, and in the process, the exit probability of each layer of the encoder and the negative slope of the LeakyReLU sublayer are set to obtain the reconstruction loss: l. the rec (x,t)=-E(xlogt);
Wherein t and x have the same size m, and the topic distribution obtained by the neural topic model is adjusted by minimizing the maximum mean difference between dirichlet distribution P, and the formula is as follows:
the overall training goals are: l = l rec (x,t)+r·λ·l MMD (t,t′)
t' is the subject distribution randomly extracted from P, k () is the information diffusion function, i, j take on values from 1 to m;
5. The topic-fused and multi-modal emergent event emotion analysis method according to claim 1, wherein the large corpus is a DBPedia dataset, and the target dataset is a dataset of netizen emotion recognition match during epidemic situations of CCIR 2020; the emotion dictionary is an emotion polarity dictionary of Taiwan university; training the neural topic model on the DBPedia data set once to finish pre-training; then, fine tuning is done on the data set of the netizen emotion recognition game during the epidemic situation of CCIR 2020.
6. The method for analyzing emergent events and emotions integrating topics and multiple modalities according to claim 1 or 5, wherein the fine tuning starts from a pre-trained model, the parameters are randomly re-initialized at the last layer of an encoder and the first layer of the decoder, the emotion value of each topic is obtained according to an emotion dictionary, and then the emotion value M of the whole comment based on the topic is obtained 1 。
7. The method for analyzing emotion of emergency event combining topic and multi-modality as claimed in claim 1, wherein in the step 2, the text and picture correlation analysis is performed by using a fusion method of BilSTM and attention mechanism, and the method comprises the following steps:
firstly, processing a text and a picture, converting the text into a text matrix by adopting a Glove method, extracting a picture label by utilizing a tool provided by a Vision Platform in a Google Cloud Platform, and expressing the picture label into a word matrix which is the same as the text;
then using two independent BilSTMs to respectively receive the picture label and the text label, and representing the picture and the text as vectors with the same dimensionality through the BilSTM;
finally, feature splicing is carried out on the image vector and the text vector to serve as input of a full connection layer, and finally the image-text correlation coefficient mu is output through a softmax layer.
8. The method for analyzing emergent events and emotions integrating topics and multimodalities according to claim 1, wherein in the step 3, based on a text classification method of a BilSTM-Attention-textCNN hybrid neural network, the text in the comment to be analyzed is mapped into a vector through a word embedding layer, and the BilSTM network is used for learning the upper expression and the lower expression of the word in the comment to be analyzed, so as to obtain a semantic vector of the current word at a deeper level; establishing an attention model, and calculating the probability weight of each word vector to enable words with larger weights to get more attention, wherein the words with more attention are often key words for classification tasks; connecting the vectors output by the attention mechanism with a pooling layer, performing k-max pooling, and reserving the first k words with larger weights; connecting textCNN network extraction feature output text emotion value
A picture classification method based on a BilSTM-Attention-CNN hybrid neural network comprises the steps of extracting labels from pictures in comments to be analyzed, representing the labels into matrixes the same as texts, extracting picture features by using the BilSTM, establishing an Attention mechanism to select the features, connecting vectors output by the Attention mechanism to a pooling layer, performing k-max pooling, and preservingLeaving the first k words with larger weight; connecting CNN network to extract characteristic output picture emotion value
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211003522.8A CN115392232A (en) | 2022-08-19 | 2022-08-19 | Topic and multi-mode fused emergency emotion analysis method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211003522.8A CN115392232A (en) | 2022-08-19 | 2022-08-19 | Topic and multi-mode fused emergency emotion analysis method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN115392232A true CN115392232A (en) | 2022-11-25 |
Family
ID=84121375
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211003522.8A Pending CN115392232A (en) | 2022-08-19 | 2022-08-19 | Topic and multi-mode fused emergency emotion analysis method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115392232A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115982473A (en) * | 2023-03-21 | 2023-04-18 | 环球数科集团有限公司 | AIGC-based public opinion analysis arrangement system |
-
2022
- 2022-08-19 CN CN202211003522.8A patent/CN115392232A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115982473A (en) * | 2023-03-21 | 2023-04-18 | 环球数科集团有限公司 | AIGC-based public opinion analysis arrangement system |
CN115982473B (en) * | 2023-03-21 | 2023-06-23 | 环球数科集团有限公司 | Public opinion analysis arrangement system based on AIGC |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113158665B (en) | Method for improving dialog text generation based on text abstract generation and bidirectional corpus generation | |
CN112860888B (en) | Attention mechanism-based bimodal emotion analysis method | |
CN107247702A (en) | A kind of text emotion analysis and processing method and system | |
CN112651448A (en) | Multi-modal emotion analysis method for social platform expression package | |
CN109829499B (en) | Image-text data fusion emotion classification method and device based on same feature space | |
CN115329779A (en) | Multi-person conversation emotion recognition method | |
CN114549850B (en) | Multi-mode image aesthetic quality evaluation method for solving modal missing problem | |
CN112800225B (en) | Microblog comment emotion classification method and system | |
CN113822340A (en) | Image-text emotion recognition method based on attention mechanism | |
CN113657115A (en) | Multi-modal Mongolian emotion analysis method based on ironic recognition and fine-grained feature fusion | |
CN114417097A (en) | Emotion prediction method and system based on time convolution and self-attention | |
CN116028846A (en) | Multi-mode emotion analysis method integrating multi-feature and attention mechanisms | |
CN114004220A (en) | Text emotion reason identification method based on CPC-ANN | |
Onita et al. | From image to text in sentiment analysis via regression and deep learning | |
CN115775349A (en) | False news detection method and device based on multi-mode fusion | |
CN115392232A (en) | Topic and multi-mode fused emergency emotion analysis method | |
CN113947074A (en) | Deep collaborative interaction emotion reason joint extraction method | |
CN113627550A (en) | Image-text emotion analysis method based on multi-mode fusion | |
CN116913278B (en) | Voice processing method, device, equipment and storage medium | |
CN113486143A (en) | User portrait generation method based on multi-level text representation and model fusion | |
CN115017900B (en) | Conversation emotion recognition method based on multi-mode multi-prejudice | |
CN114020871B (en) | Multi-mode social media emotion analysis method based on feature fusion | |
CN114443846A (en) | Classification method and device based on multi-level text abnormal composition and electronic equipment | |
CN114386412A (en) | Multi-modal named entity recognition method based on uncertainty perception | |
Chen et al. | Accelerating Semi-Supervised Text Classification by K-Way Projecting Networks |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |