CN113591462A - Bullet screen reply generation method and device and electronic equipment - Google Patents

Bullet screen reply generation method and device and electronic equipment Download PDF

Info

Publication number
CN113591462A
CN113591462A CN202110855573.2A CN202110855573A CN113591462A CN 113591462 A CN113591462 A CN 113591462A CN 202110855573 A CN202110855573 A CN 202110855573A CN 113591462 A CN113591462 A CN 113591462A
Authority
CN
China
Prior art keywords
bullet screen
vector
screen information
word
theme
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110855573.2A
Other languages
Chinese (zh)
Inventor
刘伟伟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
China Mobile Communications Group Co Ltd
MIGU Digital Media Co Ltd
MIGU Culture Technology Co Ltd
Original Assignee
China Mobile Communications Group Co Ltd
MIGU Digital Media Co Ltd
MIGU Culture Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by China Mobile Communications Group Co Ltd, MIGU Digital Media Co Ltd, MIGU Culture Technology Co Ltd filed Critical China Mobile Communications Group Co Ltd
Priority to CN202110855573.2A priority Critical patent/CN113591462A/en
Publication of CN113591462A publication Critical patent/CN113591462A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N21/00Selective content distribution, e.g. interactive television or video on demand [VOD]
    • H04N21/40Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
    • H04N21/47End-user applications
    • H04N21/488Data services, e.g. news ticker
    • H04N21/4884Data services, e.g. news ticker for displaying subtitles
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/258Heading extraction; Automatic titling; Numbering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/30Semantic analysis
    • G06F40/35Discourse or dialogue representation

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • General Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Machine Translation (AREA)

Abstract

The embodiment of the invention relates to the technical field of artificial intelligence and discloses a bullet screen reply generation method and device and electronic equipment. The method comprises the following steps: acquiring the bullet screen information; preprocessing the bullet screen information to convert the bullet screen information into an initial text vector; determining a subject matter semantic relation distribution of the initial text vector; determining a syntax dependence vector and a position relation vector between words in the bullet screen information according to the initial text vector; performing weighted calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word; and inputting the input vector and the subject semantic relation distribution into a target neural network model for prediction so as to generate a bullet screen information context for replying the bullet screen information context. Through the mode, the bullet screen recovery quality is improved.

Description

Bullet screen reply generation method and device and electronic equipment
Technical Field
The embodiment of the invention relates to the technical field of artificial intelligence, in particular to a bullet screen reply generation method and device and electronic equipment.
Background
A bullet screen is a segment of text that is displayed in synchronization with a video for user interaction with the video or between users. The user can better participate in the video communication and interaction by issuing the barrage related to the video situation.
According to the bullet screen information issued by the user, the bullet screen reply information is generated, so that the participation and immersion of the user can be effectively improved. In the related art, the generation of the bullet screen reply information is mainly based on an extraction method and an understanding generation method. However, the inventors found in the course of implementing embodiments of the present invention that: the quality of the bullet screen reply information generated in the related technology is poor, for example, the interactivity of the bullet screen reply information generated by adopting an extraction method is poor, and the bullet screen reply information generated by adopting an understanding generation method has defects in semantic logic.
Disclosure of Invention
In view of the foregoing problems, embodiments of the present invention provide a bullet screen reply generation method, device and electronic device, so as to solve the problem in the prior art that the bullet screen reply information quality is poor.
According to an aspect of the embodiments of the present invention, there is provided a bullet screen reply generation method, including:
acquiring the bullet screen information;
preprocessing the bullet screen information to convert the bullet screen information into an initial text vector;
determining a subject matter semantic relation distribution of the initial text vector;
determining a syntax dependence vector and a position relation vector between words in the bullet screen information according to the initial text vector;
performing weighted calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word;
and inputting the input vector and the subject semantic relation distribution into a target neural network for prediction so as to generate a bullet screen information context for replying the bullet screen information context.
In an optional manner, the determining, according to the initial text vector, a syntactic dependency vector and a position relationship vector between words in the bullet screen information includes:
determining the position of a current word, wherein the current word is any word vector in the initial text vectors;
traversing the position of the dependent word corresponding to the current word, and ending the traversal when traversing to the position of the core word, wherein the dependent word is the dependent word of the current word;
calculating the absolute value of the difference value between the position of the current word and the position of the dependent word;
and accumulating the absolute value of the difference value to calculate the core weight of the current word to obtain the syntactic dependency vector.
In an alternative approach, the target neural network includes an attention mechanism, a feed-forward network, and a subject word prediction model; the step of inputting the input vector and the subject semantic relation distribution into a target neural network for prediction to generate a bullet screen information context for replying the bullet screen information context comprises:
determining the theme of the initial text vector and the corresponding word distribution under the theme according to the theme semantic relationship distribution;
according to the theme and the corresponding word distribution under the theme, performing attention mechanism calculation on the input vector to obtain an interword relation vector;
after residual connection and normalization are carried out on the inter-word relation vectors, the inter-word relation vectors are input into the feedforward network, and theme semantic vectors of all words in the input vectors are obtained;
residual error connection is carried out on the theme semantic vector and the inter-word relation vector to obtain a bullet screen information sequence vector;
and inputting the bullet screen information sequence vector into the subject word prediction model to generate a bullet screen information context for replying the bullet screen information context.
In an optional manner, the bullet screen information sequence vector includes a theme above the bullet screen information, word distribution under the theme, theme probability and semantic information; the inputting the bullet screen information sequence vector into the subject term prediction model to generate a bullet screen information context for replying the bullet screen information context includes:
determining a last prediction result, wherein the last prediction result is a predicted word predicted before a current predicted word;
taking the last prediction result and the bullet screen information sequence vector as input, and obtaining the probability of the current predicted word according to a preset vocabulary model and the distribution of the subject semantic relation;
weighting and calculating the theme probability and the current predicted word probability to obtain corresponding predicted words under each theme;
and generating a bullet screen information context for replying the bullet screen information context according to the corresponding prediction words under the topics.
In an optional manner, the method further comprises:
calculating a quality value of the bullet screen information according to a preset quality value formula, wherein the quality value is used for representing the logic collocation quality, the content diversity quality and the theme quality of the bullet screen information;
and if the quality value of the bullet screen information context is larger than a preset quality value threshold, outputting the bullet screen information context.
In an alternative manner, calculating the mass value of the bullet screen information according to a preset mass value formula includes:
respectively determining a logic collocation quality value, a content diversity quality value and a theme quality value under the bullet screen information;
calculating a product of the logical collocation quality value and the content diversity quality value, and determining a sum of the product and the topic quality value as the quality value.
In an optional manner, the method further comprises:
inputting a training sample into the target neural network for training to obtain an output result; the training sample is a bullet screen dialogue text with a label; the label comprises a theme of the bullet screen dialogue text and a corresponding bullet screen dialogue text;
calculating a loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function;
adjusting parameters of the target neural network according to the loss value, re-executing the step of inputting a training sample into the target neural network for training to obtain an output result, calculating the loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function, and adjusting the parameters of the target neural network according to the loss value until the preset loss function is converged or is smaller than a preset threshold value;
the preset loss function is composed of a word prediction likelihood function and a cosine similarity distance of a context semantic vector.
According to another aspect of the embodiments of the present invention, there is provided a bullet screen reply generation apparatus, including:
the acquisition module is used for acquiring the bullet screen information;
the conversion module is used for preprocessing the bullet screen information and converting the bullet screen information into an initial text vector;
the first determining module is used for determining the subject semantic relation distribution of the initial text vector;
the second determining module is used for determining a syntactic dependency vector and a position relation vector between words in the bullet screen information according to the initial text vector;
the calculation module is used for carrying out weighting calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word;
and the prediction module is used for inputting the input vector and the subject semantic relation distribution into a target neural network for prediction so as to generate a bullet screen information context for replying the bullet screen information context.
According to another aspect of the embodiments of the present invention, there is provided an electronic device including: the system comprises a processor, a memory, a communication interface and a communication bus, wherein the processor, the memory and the communication interface complete mutual communication through the communication bus;
the memory is used for storing at least one executable instruction, and the executable instruction enables the processor to execute the operation of the bullet screen reply generation method.
According to another aspect of the embodiments of the present invention, a computer-readable storage medium is provided, where at least one executable instruction is stored in the storage medium, and when the executable instruction is executed on an electronic device, the electronic device executes the operations of the bullet screen reply generation method described above.
In the embodiment of the invention, after converting the bullet screen information into the initial text vector, the theme semantic relationship distribution of the initial text vector, and the syntax dependence vector and the position relationship vector between words in the bullet screen information can be determined; and determining an input vector of each word according to the syntactic dependency vector and the position relation vector, inputting the input vector and the subject semantic relation distribution into a target neural network for prediction, and generating a bullet screen information context for replying the bullet screen information context. The method and the device can determine the input vector of each word through the syntactic dependency vector and the position relation vector, and can more accurately mine semantic roles borne by the predicted words, so that the logic structure of the bullet screen information is more accurate; the theme characteristics above the bullet screen information can be dynamically captured by determining the theme semantic relationship distribution of the initial text vector, so that the theme characteristics below the bullet screen information are consistent with the theme characteristics above the bullet screen information, and the generation quality of the bullet screen information is further improved.
The foregoing description is only an overview of the technical solutions of the embodiments of the present invention, and the embodiments of the present invention can be implemented according to the content of the description in order to make the technical means of the embodiments of the present invention more clearly understood, and the detailed description of the present invention is provided below in order to make the foregoing and other objects, features, and advantages of the embodiments of the present invention more clearly understandable.
Drawings
The drawings are only for purposes of illustrating embodiments and are not to be construed as limiting the invention. Also, like reference numerals are used to refer to like parts throughout the drawings. In the drawings:
fig. 1 is a schematic structural diagram illustrating a content generation network provided by an embodiment of the present invention;
fig. 2 is a schematic flow chart illustrating a bullet screen reply generation method according to an embodiment of the present invention;
fig. 3 is a schematic structural diagram illustrating a bullet screen reply generation apparatus according to an embodiment of the present invention;
fig. 4 shows a schematic structural diagram of an electronic device provided in an embodiment of the present invention.
Detailed Description
Exemplary embodiments of the present invention will be described in more detail below with reference to the accompanying drawings. While exemplary embodiments of the invention are shown in the drawings, it should be understood that the invention can be embodied in various forms and should not be limited to the embodiments set forth herein.
Fig. 1 is a schematic structural diagram illustrating a content generation network according to an embodiment of the present invention. As shown in fig. 1, the content generation network is a bullet screen theme generation network constructed based on a Transformer deep learning model, and includes a semantic learning module, a theme learning module, and a bullet screen prediction module.
The semantic learning module is used for learning the main semantic logical relationship of the bullet screen based on attention mechanics by adding syntactic relationship learning codes to the bullet screen data, so that noise logical interference is reduced; in the theme learning module, by adding theme relation identification information into the bullet screen data, the theme semantic relation distribution of sentences and words is obtained through context information; and in the bullet screen prediction module, words under the subjects similar to the upper part of the bullet screen are obtained through the context to generate a conversation, so that the generated conversation is flexible and various.
Further, the embodiment of the present invention provides an evaluation feedback method, i.e., an STGE (Semantic Text generation) method, for the quality of automatically generated bullet screen content. The method evaluates the logic collocation quality, the content diversity quality and the theme quality of the bullet screen sentence, provides reference for the defects of the existing automatic evaluation method, and simultaneously provides feedback for the generation of high-quality conversation. The following describes the implementation process of the content generation network in detail with reference to specific embodiments.
Fig. 2 is a flowchart illustrating a bullet screen reply generation method according to an embodiment of the present invention, where the method is executed by an electronic device. The memory of the electronic device is used for storing at least one executable instruction, and the executable instruction enables the processor of the electronic device to execute the operation of the bullet screen reply generation method. As shown in fig. 1 and fig. 2, the method includes the following steps:
step 110: and acquiring the bullet screen information.
Wherein, the bullet screen information is generally edited and released by the user. The barrage information above may be, for example, "teacher happy! "," handset and my shape do not leave ", etc. Barrage information is generally associated with the current video above, e.g., the subject of the current video is "teacher's festival," the barrage information may be, e.g., "teacher's festival happy! When the theme of the current video is "mobile phone", the bullet screen information may be "mobile phone and my movie do not leave" above, for example. Since the theme is dynamically changed during the video playing process, the barrage information published by the user generally has real-time property, namely, is related to the current video theme.
Step 120: and preprocessing the bullet screen information and converting the bullet screen information into an initial text vector.
When the bullet screen information is preprocessed, the bullet screen information can be subjected to word segmentation, then word vectors of all words are calculated, and initial text vectors corresponding to the bullet screen information can be obtained according to the word vectors of all the words. For example, if the bullet screen information is "useless words are not much, three links are worship", and the word segmentation can be performed to obtain "useless words/not much/words/,/three links/is/worship/".
Step 130: and determining the subject semantic relation distribution of the initial text vector.
The theme semantic relation distribution of the bullet screen dialogue sequence can be trained in advance through n-layer coding learning. The theme semantic relationship distribution of the pre-training barrage conversation sequence comprises theme semantic relationship distribution corresponding to each word in the barrage conversation sequence, wherein one word may correspond to a plurality of themes, and each theme corresponds to a theme probability.
The pre-trained topic vector Z may be represented as { Z1, Z2, z3... zm }, the word vector w corresponding to the initial text vector may be represented as { d1, d2, d3... dn }, and the topic semantic relationship distribution of the initial text vector may be determined by the topic semantic relationship distribution corresponding to the word vector w in the initial text vector. Further, the distribution of the semantic relationship of the topics corresponding to the word vector w can be determined by constructing a relationship matrix of the word vector and the topic vector Z. In the initial text vector, the topic probability of the word corresponding to the word vector w can be determined by the following formula.
Figure BDA0003183987730000071
Wherein, p (z)k| w) represents the topic probability of the word corresponding to the word vector w, zkRepresenting the topic to which the word vector w belongs, m representing the total dimension of the topic vector Z, n representing the total dimension of the word vector w, p (Z) can be obtained by the above calculation formulak| w) is expressed as the semantic distance of the word vector from the corresponding topic vector.
Therefore, topic similarity detection is carried out on each word vector in the initial text vector and topic semantic relation distribution of the pre-training barrage dialogue sequence, so that topic semantic relation distribution corresponding to the initial text vector is obtained, and the topic semantic relation distribution comprises topics and topic probabilities corresponding to each word.
Step 140: and determining a syntactic dependency vector and a position relation vector among the words in the bullet screen information according to the initial text vector.
The syntax dependence vector is used for representing the semantic logic relationship of each word in the bullet screen information, and the position relationship vector is used for representing the position relationship of each word in the bullet screen information. The part of speech analysis can be carried out on the bullet screen information, and the semantic relation of each participle is analyzed. For example, "the word is not much spoken, the word" three links are worship "and after the part of speech analysis:
(ROOT (IP (NP (NP (NR WASTE))) (NP (NN NOT))) (VP (VV so (PU'))) (IP (VP (QP (CD Tri)))) (VP (VC so (NP (NN Jing)))))))))))))))))
Where ROOT denotes a sentence of a text to be processed, IP denotes a simple clause, NP denotes a noun phrase, NR denotes a proper noun, NN denotes a common noun, VP denotes a verb phrase, VV denotes a verb, PU denotes a sentence break such as a period, question mark, exclamation mark, QP denotes a quantifier phrase, and CD denotes a radix word. The result of analyzing the semantic relation of each participle is as follows:
nmod, assmod (no more than-2, waste word-1), nsubj (say-3, no more than-2), ROOT (ROOT-0, say-3), punct (say-3, -4), conj (Jing-7, san Lian-5), cop (Jing-7, is-6), ccomp (say-3, Jing-7).
Wherein nmod (non compound modifier) represents compound noun modification, assmod (associative modifier) represents associated modification, nsubj (non subject) represents noun subject, root represents root node, designation represents punctuation symbol, conj (join) represents connective, cop (copula) represents systematic verb, and ccomp represents subordinate clause complement.
The syntax dependency vector can be calculated by the following equation, for example.
Figure BDA0003183987730000081
In the above formula, p represents the position of the current word, pkdThe position of the dependent word is determined, r is the position of the core word, i (the value range is 0-d) represents the dimension of the current word, and d represents the total dimension. SE (p, i) represents a syntactic dependency vector of a word at the p-th position in the sequence of words { w1, w 2.. wn }. Further, when determining a syntactic dependency vector between words in the bullet screen information according to the initial text vector, the position of the current word can be determined first, and the current word can be any word vector in the initial text vector; traversing the position of a dependent word corresponding to the current word, and ending the traversal when the position of the core word is traversed, wherein the dependent word is the dependent word of the current word; calculating the absolute difference between the current word position and the dependent word positionA value; and accumulating the absolute values of the difference values to calculate the core weight of the current word so as to obtain the syntactic dependency vector corresponding to the current word. For example, in "do nothing, three links are worship", the core word is "say", the stem of the whole sentence surrounding the core word is "do nothing, worship", and for "do nothing", the pk position is 2, pkdPosition 3. By coding the syntactic relation through the formula, the core relation words such as the chief and the predicate guest can be focused on, so that the influence of other meaningless noises is reduced, the attention learning of the subject characteristics is focused, and the accurate identification of the logic relation among the words is improved.
Further, the positional relationship vector can be calculated by using the following equation, for example.
Figure BDA0003183987730000082
Where pos represents the location of the current word, dmodel represents the total dimension, and i represents the dimension of the location relationship vector.
Step 150: and performing weighted calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word.
And performing weighted calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word, so that the semantic relationship characteristic and the position relationship characteristic of each word can be represented by the input vector of each word. In the embodiment of the present invention, the syntactic dependency relationship vector and the position relationship vector corresponding to each word may be directly added or added after setting different weights respectively to obtain the input vector. The embodiment of the present invention does not specifically limit how to weight the specific form, and those skilled in the art can set the weight accordingly according to the specific scenario.
Step 160: and inputting the input vector and the subject semantic relation distribution into a target neural network for prediction so as to generate a bullet screen information context for replying the bullet screen information context.
Wherein, the target neural network is a neural network generated through training. The target neural network may be based on, for example, a Transformer model, which is a natural language processing model developed by ***.
In an alternative approach, the target neural network includes an attention mechanism, a feed-forward network, and a subject word prediction model. When the input vector and the distribution of the subject semantic relationship are input into a target neural network for prediction, the subject of an initial text vector and the distribution of words corresponding to the subject can be determined according to the distribution of the subject semantic relationship, attention mechanism calculation is carried out on the input vector according to the subject and the distribution of the words corresponding to the subject to obtain an inter-word relationship vector, the inter-word relationship vector is subjected to residual connection and normalization, and then the inter-word relationship vector is input into a feed-forward network to obtain the subject semantic vector of each word in the input vector; residual error connection is carried out on the theme semantic vector and the interword relation vector to obtain a bullet screen information sequence vector; and inputting the bullet screen information sequence vector into a subject word prediction model to generate a bullet screen information context for replying the bullet screen information context.
The forward propagation may be performed through a feedforward neural network, and the forward propagation process may use, for example, a GELU (Gaussian Error Linear Units) activation function. The GELU activation function may be calculated using the following equation, for example.
Figure BDA0003183987730000091
In an optional manner, the bullet screen information sequence vector includes a topic above the bullet screen information, a word distribution under the topic, a topic probability, and semantic information. When the bullet screen information sequence vector is input into the subject word prediction model, a last prediction result can be determined firstly, wherein the last prediction result is a predicted word predicted before a current predicted word; and taking the last prediction result and the bullet screen information sequence vector as input, and obtaining the probability of the current predicted word according to the preset vocabulary model and the theme semantic relationship distribution. The preset vocabulary model may be manually preset for encoding words, thereby establishing a unique correspondence between words and corresponding codes. Further, carrying out weighted calculation on the theme probability and the current predicted word probability to obtain corresponding predicted words under each theme; and obtaining a plurality of pieces of predicted bullet screen reply information and corresponding probabilities according to corresponding predicted words under each theme, and determining the predicted bullet screen reply information with the highest probability as the bullet screen information context.
Further, a word prediction function may be used to generate the corresponding predicted word under each topic, and the word prediction function may be calculated using the following formula, for example.
Figure BDA0003183987730000101
Wherein V represents a preset vocabulary model, K represents topic semantic relation distribution, V and K are subsets of V and K respectively, s represents a topic vector above bullet screen information, yi-1Representing the last prediction, w and b are the network parameters used to represent the coefficients and bias terms, respectively, p (y)i) As a function of word prediction, wTDenotes the transposition of W, WSV、Wsk、Wyv、WykAnd WxKRespectively for representing the components of w, x representing the word vector above the bullet screen information, bvAnd bkRepresenting the respective components of b, respectively.
It should be noted that, after the bullet screen information context for replying to the bullet screen information context is generated according to the above embodiment, the bullet screen information context may be further detected to determine the quality value of the bullet screen information context, and the quality value of the bullet screen information context is compared with the preset quality value threshold.
In an alternative manner, when determining the quality value of the bullet screen information context, the bullet screen information context is first divided into a plurality of logic detection units, for example, a punctuation may be added to the bullet screen information context according to a preset step length, and the logic detection units are divided according to the punctuation. And adopting a jumping mechanism to carry out logic collocation detection on different logic detection units, and determining a logic collocation quality value under the bullet screen information according to a logic collocation detection result. Furthermore, weighting can be carried out on the theme vector of each word in the bullet screen information to determine the theme vector in the bullet screen information, calculating the cosine similarity between the theme vector in the bullet screen information and the theme vector in the bullet screen information, and determining the cosine similarity as the theme quality value in the bullet screen information. Further, jump combination of different sliding windows can be performed on the word sequences under the bullet screen information, for example, word combination (w1, 1, w2) represents word combination w1w2 with distance of 1, and the logical dependency relationship of the word combination is calculated, for example, (w1, nsubj, 1, w2) represents the logical dependency relationship of word combination w1w2, and the logical relationship between words in the training word set C and the word collocation combination are combined to count the logical and specificity of the word combination between different distances in the bullet screen information, so that the logical collocation quality value and the content diversity quality value in the bullet screen information are calculated.
In an optional mode, the quality value of the bullet screen information context can be calculated according to a preset quality value formula, the quality value is used for representing the logic collocation quality, the content diversity quality and the theme quality of the bullet screen information context, and if the quality value of the bullet screen information context is larger than a preset quality value threshold, the bullet screen information context is output.
In an alternative mode, when the quality value of the bullet screen information context is calculated according to the preset quality value formula, the logic collocation quality value, the content diversity quality value and the theme quality value of the bullet screen information context can be respectively determined, the product of the logic collocation quality value and the content diversity quality value is calculated, and the sum of the product and the theme quality value is determined as the quality value of the bullet screen information context.
The preset quality value formula may be, for example, the following formula:
Figure BDA0003183987730000111
wherein the content of the first and second substances,
Figure BDA0003183987730000112
a theme quality value representing the context of the bullet screen information,
Figure BDA0003183987730000113
a topic vector representing the context of the bullet screen information,
Figure BDA0003183987730000114
the theme vector representing the bullet screen information is used for representing the theme quality value through the cosine similarity between the theme vector in the bullet screen information and the theme vector in the bullet screen information;
Figure BDA0003183987730000115
a logical collocation quality value is represented below the bullet screen information,
Figure BDA0003183987730000116
representing the content diversity quality value of the bullet screen information, r representing the set of reply words, C representing the set of training words, (w)m,i,wn) Denotes w at an interval im wnPhrase, count is used to count w without deduplicationm wnNumber of phrases, uniq for statistical deduplication wm wnThe number of phrases, m and n are self-defined parameters.
It should be noted that before inputting the input vector and the subject semantic relationship distribution into the target neural network for prediction, the target neural network needs to be trained in advance.
In an optional mode, inputting a training sample into a target neural network for training to obtain an output result, wherein the training sample is a bullet screen dialogue upper text with a label, and the label comprises a theme of the bullet screen dialogue upper text and a corresponding bullet screen dialogue lower text; calculating a loss value of the target neural network according to an output result of the target neural network, the bullet screen dialogue context and a preset loss function, adjusting parameters of the target neural network according to the loss value of the target neural network, training the target neural network again according to the training sample, and adjusting the parameters of the target neural network according to the loss value of the target neural network until the preset loss function is converged or is smaller than a preset threshold value.
In an alternative, the predetermined loss function may be formed by a cosine similarity distance between the word prediction likelihood function and the context semantic vector. For example, the preset loss function may be, for example:
Figure BDA0003183987730000121
wherein logP (x)i|xi-k,.....xi-1Theta) is a word prediction likelihood function,
Figure BDA0003183987730000122
the cosine similarity distance of the context semantic vector,
Figure BDA0003183987730000123
the above theme vector of the bullet screen is shown,
Figure BDA0003183987730000124
representing the bullet screen context topic vector. Alpha and beta are adjustable parameters, theta is a parameter to be estimated, and k and m are self-defined parameters.
In the embodiment of the invention, after converting the bullet screen information into the initial text vector, the theme semantic relationship distribution of the initial text vector, and the syntax dependence vector and the position relationship vector between words in the bullet screen information can be determined; and determining an input vector of each word according to the syntactic dependency vector and the position relation vector, inputting the input vector and the subject semantic relation distribution into a target neural network for prediction, and generating a bullet screen information context for replying the bullet screen information context. The method and the device can determine the input vector of each word through the syntactic dependency vector and the position relation vector, and can more accurately mine semantic roles borne by the predicted words, so that the logic structure of the bullet screen information is more accurate; the theme characteristics above the bullet screen information can be dynamically captured by determining the theme semantic relationship distribution of the initial text vector, so that the theme characteristics below the bullet screen information are consistent with the theme characteristics above the bullet screen information, and the generation quality of the bullet screen information is further improved.
Fig. 3 shows a schematic structural diagram of a bullet screen generating device according to an embodiment of the present invention. As shown in fig. 3, the apparatus 300 includes: an acquisition module 310, a conversion module 320, a first determination module 330, a second determination module 340, a calculation module 350, and a prediction module 360.
The obtaining module 310 is configured to obtain the bullet screen information;
the conversion module 320 is configured to preprocess the bullet screen information to convert the bullet screen information into an initial text vector;
a first determining module 330, configured to determine a subject semantic relationship distribution of the initial text vector;
a second determining module 340, configured to determine, according to the initial text vector, a syntactic dependency vector and a position relationship vector between words in the bullet screen information;
a calculating module 350, configured to perform weighted calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word;
and the prediction module 360 is configured to input the input vector and the subject semantic relationship distribution into a target neural network for prediction, so as to generate a bullet screen information context for replying the bullet screen information context.
In an alternative manner, the second determining module 340 is configured to:
determining the position of a current word, wherein the current word is any word vector in the initial text vectors;
traversing the position of the dependent word corresponding to the current word, and ending the traversal when traversing to the position of the core word, wherein the dependent word is the dependent word of the current word;
calculating the absolute value of the difference value between the position of the current word and the position of the dependent word;
and accumulating the absolute value of the difference value to calculate the core weight of the current word to obtain the syntactic dependency vector.
In an alternative approach, the target neural network includes an attention mechanism, a feed-forward network, and a subject word prediction model; the prediction module 360 is to:
determining the theme of the initial text vector and the corresponding word distribution under the theme according to the theme semantic relationship distribution;
according to the theme and the corresponding word distribution under the theme, performing attention mechanism calculation on the input vector to obtain an interword relation vector;
after residual connection and normalization are carried out on the inter-word relation vectors, the inter-word relation vectors are input into the feedforward network, and theme semantic vectors of all words in the input vectors are obtained;
residual error connection is carried out on the theme semantic vector and the inter-word relation vector to obtain a bullet screen information sequence vector;
and inputting the bullet screen information sequence vector into the subject word prediction model to generate a bullet screen information context for replying the bullet screen information context.
In an optional manner, the bullet screen information sequence vector includes a theme above the bullet screen information, word distribution under the theme, theme probability and semantic information; the prediction module 360 is to:
determining a last prediction result, wherein the last prediction result is a predicted word predicted before a current predicted word;
taking the last prediction result and the bullet screen information sequence vector as input, and obtaining the probability of the current predicted word according to a preset vocabulary model and the distribution of the subject semantic relation;
weighting and calculating the theme probability and the current predicted word probability to obtain corresponding predicted words under each theme;
and generating a bullet screen information context for replying the bullet screen information context according to the corresponding prediction words under the topics.
In an optional manner, the apparatus 300 further comprises an output module for:
calculating a quality value of the bullet screen information according to a preset quality value formula, wherein the quality value is used for representing the logic collocation quality, the content diversity quality and the theme quality of the bullet screen information;
and if the quality value of the bullet screen information context is larger than a preset quality value threshold, outputting the bullet screen information context.
In an alternative form, the output module is configured to:
respectively determining a logic collocation quality value, a content diversity quality value and a theme quality value under the bullet screen information;
calculating a product of the logical collocation quality value and the content diversity quality value, and determining a sum of the product and the topic quality value as the quality value.
In an optional manner, the apparatus 300 further comprises a training module for:
inputting a training sample into the target neural network for training to obtain an output result, wherein the training sample is a bullet screen dialogue text with a label, and the label comprises a theme of the bullet screen dialogue text and a corresponding bullet screen dialogue text;
calculating a loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function;
adjusting parameters of the target neural network according to the loss value, re-executing the step of inputting a training sample into the target neural network for training to obtain an output result, calculating the loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function, and adjusting the parameters of the target neural network according to the loss value until the preset loss function is converged or is smaller than a preset threshold value;
the preset loss function is composed of a word prediction likelihood function and a cosine similarity distance of a context semantic vector.
In the embodiment of the invention, after converting the bullet screen information into the initial text vector, the theme semantic relationship distribution of the initial text vector, and the syntax dependence vector and the position relationship vector between words in the bullet screen information can be determined; and determining an input vector of each word according to the syntactic dependency vector and the position relation vector, inputting the input vector and the subject semantic relation distribution into a target neural network for prediction, and generating a bullet screen information context for replying the bullet screen information context. The method and the device can determine the input vector of each word through the syntactic dependency vector and the position relation vector, and can more accurately mine semantic roles borne by the predicted words, so that the logic structure of the bullet screen information is more accurate; the theme characteristics above the bullet screen information can be dynamically captured by determining the theme semantic relationship distribution of the initial text vector, so that the theme characteristics below the bullet screen information are consistent with the theme characteristics above the bullet screen information, and the generation quality of the bullet screen information is further improved.
Fig. 4 is a schematic structural diagram of an electronic device according to an embodiment of the present invention, and the specific embodiment of the present invention does not limit the specific implementation of the electronic device.
As shown in fig. 4, the electronic device may include: a processor (processor)402, a Communications Interface 404, a memory 406, and a Communications bus 408.
Wherein: the processor 402, communication interface 404, and memory 406 communicate with each other via a communication bus 408. A communication interface 404 for communicating with network elements of other devices, such as clients or other servers. The processor 402 is configured to execute the program 410, and may specifically execute the relevant steps in the foregoing bullet screen generation method embodiment.
In particular, program 410 may include program code comprising computer-executable instructions.
The processor 402 may be a central processing unit CPU or an application Specific Integrated circuit asic or one or more Integrated circuits configured to implement embodiments of the present invention. The electronic device comprises one or more processors, which can be the same type of processor, such as one or more CPUs; or may be different types of processors such as one or more CPUs and one or more ASICs.
And a memory 406 for storing a program 410. Memory 406 may comprise high-speed RAM memory, and may also include non-volatile memory (non-volatile memory), such as at least one disk memory.
The program 410 may specifically be invoked by the processor 402 to cause the electronic device to perform the following operations:
acquiring the bullet screen information;
preprocessing the bullet screen information to convert the bullet screen information into an initial text vector;
determining a subject matter semantic relation distribution of the initial text vector;
determining a syntax dependence vector and a position relation vector between words in the bullet screen information according to the initial text vector;
performing weighted calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word;
and inputting the input vector and the subject semantic relation distribution into a target neural network for prediction so as to generate a bullet screen information context for replying the bullet screen information context.
In an alternative, the program 410 is invoked by the processor 402 to cause the electronic device to perform the following operations:
determining the position of a current word, wherein the current word is any word vector in the initial text vectors;
traversing the position of the dependent word corresponding to the current word, and ending the traversal when traversing to the position of the core word, wherein the dependent word is the dependent word of the current word;
calculating the absolute value of the difference value between the position of the current word and the position of the dependent word;
and accumulating the absolute value of the difference value to calculate the core weight of the current word to obtain the syntactic dependency vector.
In an alternative approach, the target neural network includes an attention mechanism, a feed-forward network, and a subject word prediction model; the program 410 is invoked by the processor 402 to cause the electronic device to perform the following operations:
determining the theme of the initial text vector and the corresponding word distribution under the theme according to the theme semantic relationship distribution;
according to the theme and the corresponding word distribution under the theme, performing attention mechanism calculation on the input vector to obtain an interword relation vector;
after residual connection and normalization are carried out on the inter-word relation vectors, the inter-word relation vectors are input into the feedforward network, and theme semantic vectors of all words in the input vectors are obtained;
residual error connection is carried out on the theme semantic vector and the inter-word relation vector to obtain a bullet screen information sequence vector;
and inputting the bullet screen information sequence vector into the subject word prediction model to generate a bullet screen information context for replying the bullet screen information context.
In an optional manner, the bullet screen information sequence vector includes a theme above the bullet screen information, word distribution under the theme, theme probability and semantic information; the program 410 is invoked by the processor 402 to cause the electronic device to perform the following operations:
determining a last prediction result, wherein the last prediction result is a predicted word predicted before a current predicted word;
taking the last prediction result and the bullet screen information sequence vector as input, and obtaining the probability of the current predicted word according to a preset vocabulary model and the distribution of the subject semantic relation;
weighting and calculating the theme probability and the current predicted word probability to obtain corresponding predicted words under each theme;
and generating a bullet screen information context for replying the bullet screen information context according to the corresponding prediction words under the topics.
In an alternative, the program 410 is invoked by the processor 402 to cause the electronic device to perform the following operations:
calculating a quality value of the bullet screen information according to a preset quality value formula, wherein the quality value is used for representing the logic collocation quality, the content diversity quality and the theme quality of the bullet screen information;
and if the quality value of the bullet screen information context is larger than a preset quality value threshold, outputting the bullet screen information context.
In an alternative, the program 410 is invoked by the processor 402 to cause the electronic device to perform the following operations:
respectively determining a logic collocation quality value, a content diversity quality value and a theme quality value under the bullet screen information;
calculating a product of the logical collocation quality value and the content diversity quality value, and determining a sum of the product and the topic quality value as the quality value.
In an alternative, the program 410 is invoked by the processor 402 to cause the electronic device to perform the following operations:
inputting a training sample into the target neural network for training to obtain an output result; the training sample is a bullet screen dialogue text with a label; the label comprises a theme of the bullet screen dialogue text and a corresponding bullet screen dialogue text;
calculating a loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function;
adjusting parameters of the target neural network according to the loss value, re-executing the step of inputting a training sample into the target neural network for training to obtain an output result, calculating the loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function, and adjusting the parameters of the target neural network according to the loss value until the preset loss function is converged or is smaller than a preset threshold value;
the preset loss function is composed of a word prediction likelihood function and a cosine similarity distance of a context semantic vector.
In the embodiment of the invention, after converting the bullet screen information into the initial text vector, the theme semantic relationship distribution of the initial text vector, and the syntax dependence vector and the position relationship vector between words in the bullet screen information can be determined; and determining an input vector of each word according to the syntactic dependency vector and the position relation vector, inputting the input vector and the subject semantic relation distribution into a target neural network for prediction, and generating a bullet screen information context for replying the bullet screen information context. The method and the device can determine the input vector of each word through the syntactic dependency vector and the position relation vector, and can more accurately mine semantic roles borne by the predicted words, so that the logic structure of the bullet screen information is more accurate; the theme characteristics above the bullet screen information can be dynamically captured by determining the theme semantic relationship distribution of the initial text vector, so that the theme characteristics below the bullet screen information are consistent with the theme characteristics above the bullet screen information, and the generation quality of the bullet screen information is further improved.
The embodiment of the invention provides a computer-readable storage medium, wherein at least one executable instruction is stored in the storage medium, and when the executable instruction runs on an electronic device, the electronic device is enabled to execute the bullet screen reply generation method in any method embodiment.
The embodiment of the invention provides a bullet screen generating device which is used for executing the bullet screen reply generating method.
Embodiments of the present invention provide a computer program, where the computer program can be called by a processor to enable an electronic device to execute a bullet screen reply generation method in any of the above method embodiments.
An embodiment of the present invention provides a computer program product, where the computer program product includes a computer program stored on a computer-readable storage medium, and the computer program includes program instructions, when the program instructions are run on a computer, the computer is caused to execute the bullet screen reply generation method in any of the above method embodiments.
The algorithms or displays presented herein are not inherently related to any particular computer, virtual system, or other apparatus. Various general purpose systems may also be used with the teachings herein. The required structure for constructing such a system will be apparent from the description above. In addition, embodiments of the present invention are not directed to any particular programming language. It is appreciated that a variety of programming languages may be used to implement the teachings of the present invention as described herein, and any descriptions of specific languages are provided above to disclose the best mode of the invention.
In the description provided herein, numerous specific details are set forth. It is understood, however, that embodiments of the invention may be practiced without these specific details. In some instances, well-known methods, structures and techniques have not been shown in detail in order not to obscure an understanding of this description.
Similarly, it should be appreciated that in the foregoing description of exemplary embodiments of the invention, various features of the embodiments of the invention are sometimes grouped together in a single embodiment, figure, or description thereof for the purpose of streamlining the invention and aiding in the understanding of one or more of the various inventive aspects. However, the disclosed method should not be interpreted as reflecting an intention that: that the invention as claimed requires more features than are expressly recited in each claim.
Those skilled in the art will appreciate that the modules in the device in an embodiment may be adaptively changed and disposed in one or more devices different from the embodiment. The modules or units or components of the embodiments may be combined into one module or unit or component, and may be divided into a plurality of sub-modules or sub-units or sub-components. All of the features disclosed in this specification (including any accompanying claims, abstract and drawings), and all of the processes or elements of any method or apparatus so disclosed, may be combined in any combination, except combinations where at least some of such features and/or processes or elements are mutually exclusive. Each feature disclosed in this specification (including any accompanying claims, abstract and drawings) may be replaced by alternative features serving the same, equivalent or similar purpose, unless expressly stated otherwise.
It should be noted that the above-mentioned embodiments illustrate rather than limit the invention, and that those skilled in the art will be able to design alternative embodiments without departing from the scope of the appended claims. In the claims, any reference signs placed between parentheses shall not be construed as limiting the claim. The word "comprising" does not exclude the presence of elements or steps not listed in a claim. The word "a" or "an" preceding an element does not exclude the presence of a plurality of such elements. The invention may be implemented by means of hardware comprising several distinct elements, and by means of a suitably programmed computer. In the unit claims enumerating several means, several of these means may be embodied by one and the same item of hardware. The usage of the words first, second and third, etcetera do not indicate any ordering. These words may be interpreted as names. The steps in the above embodiments should not be construed as limiting the order of execution unless specified otherwise.

Claims (10)

1. A bullet screen reply generation method is characterized by comprising the following steps:
acquiring the bullet screen information;
preprocessing the bullet screen information to convert the bullet screen information into an initial text vector;
determining a subject matter semantic relation distribution of the initial text vector;
determining a syntax dependence vector and a position relation vector between words in the bullet screen information according to the initial text vector;
performing weighted calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word;
and inputting the input vector and the subject semantic relation distribution into a target neural network for prediction so as to generate a bullet screen information context for replying the bullet screen information context.
2. The method of claim 1, wherein determining the syntactic dependency vector and the position relationship vector between words in the bullet screen information according to the initial text vector comprises:
determining the position of a current word, wherein the current word is any word vector in the initial text vectors;
traversing the position of the dependent word corresponding to the current word, and ending the traversal when traversing to the position of the core word, wherein the dependent word is the dependent word of the current word;
calculating the absolute value of the difference value between the position of the current word and the position of the dependent word;
and accumulating the absolute value of the difference value to calculate the core weight of the current word to obtain the syntactic dependency vector.
3. The method of claim 1, wherein the target neural network comprises an attention mechanism, a feed forward network, and a subject term prediction model; the step of inputting the input vector and the subject semantic relation distribution into a target neural network for prediction to generate a bullet screen information context for replying the bullet screen information context comprises:
determining the theme of the initial text vector and the corresponding word distribution under the theme according to the theme semantic relationship distribution;
according to the theme and the corresponding word distribution under the theme, performing attention mechanism calculation on the input vector to obtain an interword relation vector;
after residual connection and normalization are carried out on the inter-word relation vectors, the inter-word relation vectors are input into the feedforward network, and theme semantic vectors of all words in the input vectors are obtained;
residual error connection is carried out on the theme semantic vector and the inter-word relation vector to obtain a bullet screen information sequence vector;
and inputting the bullet screen information sequence vector into the subject word prediction model to generate a bullet screen information context for replying the bullet screen information context.
4. The method of claim 3, wherein the barrage information sequence vector comprises a topic above the barrage information, a word distribution under the topic, a topic probability, and semantic information; the inputting the bullet screen information sequence vector into the subject term prediction model to generate a bullet screen information context for replying the bullet screen information context includes:
determining a last prediction result, wherein the last prediction result is a predicted word predicted before a current predicted word;
taking the last prediction result and the bullet screen information sequence vector as input, and obtaining the probability of the current predicted word according to a preset vocabulary model and the distribution of the subject semantic relation;
weighting and calculating the theme probability and the current predicted word probability to obtain corresponding predicted words under each theme;
and generating a bullet screen information context for replying the bullet screen information context according to the corresponding prediction words under the topics.
5. The method of claim 1, further comprising:
calculating a quality value of the bullet screen information context according to a preset quality value formula, wherein the quality value is used for representing the logic collocation quality, the content diversity quality and the theme quality of the bullet screen information context;
and if the quality value of the bullet screen information context is larger than a preset quality value threshold, outputting the bullet screen information context.
6. The method of claim 5, wherein calculating a quality value for the bullet screen information context according to a preset quality value formula comprises:
respectively determining a logic collocation quality value, a content diversity quality value and a theme quality value under the bullet screen information;
calculating a product of the logical collocation quality value and the content diversity quality value, and determining a sum of the product and the topic quality value as the quality value.
7. The method of claim 1, further comprising:
inputting a training sample into the target neural network for training to obtain an output result; the training sample is a bullet screen dialogue text with a label; the label comprises a theme of the bullet screen dialogue text and a corresponding bullet screen dialogue text;
calculating a loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function;
adjusting parameters of the target neural network according to the loss value, re-executing the step of inputting a training sample into the target neural network for training to obtain an output result, calculating the loss value of the target neural network according to the output result, the bullet screen dialog context and a preset loss function, and adjusting the parameters of the target neural network according to the loss value until the preset loss function is converged or is smaller than a preset threshold value;
the preset loss function is composed of a word prediction likelihood function and a cosine similarity distance of a context semantic vector.
8. A bullet screen reply generation apparatus, characterized in that the apparatus comprises:
the acquisition module is used for acquiring the bullet screen information;
the conversion module is used for preprocessing the bullet screen information and converting the bullet screen information into an initial text vector;
the first determining module is used for determining the subject semantic relation distribution of the initial text vector;
the second determining module is used for determining a syntactic dependency vector and a position relation vector between words in the bullet screen information according to the initial text vector;
the calculation module is used for carrying out weighting calculation on the syntactic dependency relationship vector and the position relationship vector to obtain an input vector of each word;
and the prediction module is used for inputting the input vector and the subject semantic relation distribution into a target neural network for prediction so as to generate a bullet screen information context for replying the bullet screen information context.
9. An electronic device, comprising: the system comprises a processor, a memory, a communication interface and a communication bus, wherein the processor, the memory and the communication interface complete mutual communication through the communication bus;
the memory is used for storing at least one executable instruction, and the executable instruction causes the processor to execute the operation of the bullet screen reply generation method according to any one of claims 1 to 7.
10. A computer-readable storage medium, wherein the storage medium has at least one executable instruction stored therein, and when the executable instruction is executed on an electronic device, the electronic device is caused to perform the operations of the bullet screen reply generation method according to any one of claims 1 to 7.
CN202110855573.2A 2021-07-28 2021-07-28 Bullet screen reply generation method and device and electronic equipment Pending CN113591462A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110855573.2A CN113591462A (en) 2021-07-28 2021-07-28 Bullet screen reply generation method and device and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110855573.2A CN113591462A (en) 2021-07-28 2021-07-28 Bullet screen reply generation method and device and electronic equipment

Publications (1)

Publication Number Publication Date
CN113591462A true CN113591462A (en) 2021-11-02

Family

ID=78251177

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110855573.2A Pending CN113591462A (en) 2021-07-28 2021-07-28 Bullet screen reply generation method and device and electronic equipment

Country Status (1)

Country Link
CN (1) CN113591462A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114548115A (en) * 2022-02-23 2022-05-27 北京三快在线科技有限公司 Method and device for explaining compound nouns and electronic equipment
CN114942986A (en) * 2022-06-21 2022-08-26 平安科技(深圳)有限公司 Text generation method and device, computer equipment and computer readable storage medium
WO2023108981A1 (en) * 2021-12-15 2023-06-22 平安科技(深圳)有限公司 Method and apparatus for training text generation model, and storage medium and computer device
CN117521680A (en) * 2024-01-04 2024-02-06 深圳和成视讯科技有限公司 Online translation method of recorder and recorder

Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106101756A (en) * 2016-06-07 2016-11-09 腾讯科技(北京)有限公司 Barrage display packing, barrage adding method, Apparatus and system
CN107229610A (en) * 2017-03-17 2017-10-03 咪咕数字传媒有限公司 The analysis method and device of a kind of affection data
CN109408641A (en) * 2018-11-22 2019-03-01 山东工商学院 It is a kind of based on have supervision topic model file classification method and system
CN110035325A (en) * 2019-04-19 2019-07-19 广州虎牙信息科技有限公司 Barrage answering method, barrage return mechanism and live streaming equipment
CN110850983A (en) * 2019-11-13 2020-02-28 腾讯科技(深圳)有限公司 Virtual object control method and device in video live broadcast and storage medium
CN110895656A (en) * 2018-09-13 2020-03-20 武汉斗鱼网络科技有限公司 Text similarity calculation method and device, electronic equipment and storage medium
CN111241295A (en) * 2020-01-03 2020-06-05 浙江大学 Knowledge map relation data extraction method based on semantic syntax interactive network
US20200294507A1 (en) * 2019-03-11 2020-09-17 Adobe Inc. Pose-invariant Visual Speech Recognition Using A Single View Input
CN111783474A (en) * 2020-07-16 2020-10-16 厦门市美亚柏科信息股份有限公司 Comment text viewpoint information processing method and device and storage medium
US20210200957A1 (en) * 2019-12-27 2021-07-01 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and apparatus for generating dialogue model

Patent Citations (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106101756A (en) * 2016-06-07 2016-11-09 腾讯科技(北京)有限公司 Barrage display packing, barrage adding method, Apparatus and system
CN107229610A (en) * 2017-03-17 2017-10-03 咪咕数字传媒有限公司 The analysis method and device of a kind of affection data
CN110895656A (en) * 2018-09-13 2020-03-20 武汉斗鱼网络科技有限公司 Text similarity calculation method and device, electronic equipment and storage medium
CN109408641A (en) * 2018-11-22 2019-03-01 山东工商学院 It is a kind of based on have supervision topic model file classification method and system
US20200294507A1 (en) * 2019-03-11 2020-09-17 Adobe Inc. Pose-invariant Visual Speech Recognition Using A Single View Input
CN110035325A (en) * 2019-04-19 2019-07-19 广州虎牙信息科技有限公司 Barrage answering method, barrage return mechanism and live streaming equipment
CN110850983A (en) * 2019-11-13 2020-02-28 腾讯科技(深圳)有限公司 Virtual object control method and device in video live broadcast and storage medium
US20210200957A1 (en) * 2019-12-27 2021-07-01 Beijing Baidu Netcom Science And Technology Co., Ltd. Method and apparatus for generating dialogue model
CN111241295A (en) * 2020-01-03 2020-06-05 浙江大学 Knowledge map relation data extraction method based on semantic syntax interactive network
CN111783474A (en) * 2020-07-16 2020-10-16 厦门市美亚柏科信息股份有限公司 Comment text viewpoint information processing method and device and storage medium

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
李卫疆;漆芳;: "基于多通道双向长短期记忆网络的情感分析", 《中文信息学报》, no. 12, 15 December 2019 (2019-12-15) *

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2023108981A1 (en) * 2021-12-15 2023-06-22 平安科技(深圳)有限公司 Method and apparatus for training text generation model, and storage medium and computer device
CN114548115A (en) * 2022-02-23 2022-05-27 北京三快在线科技有限公司 Method and device for explaining compound nouns and electronic equipment
CN114548115B (en) * 2022-02-23 2023-01-06 北京三快在线科技有限公司 Method and device for explaining compound nouns and electronic equipment
CN114942986A (en) * 2022-06-21 2022-08-26 平安科技(深圳)有限公司 Text generation method and device, computer equipment and computer readable storage medium
CN114942986B (en) * 2022-06-21 2024-03-19 平安科技(深圳)有限公司 Text generation method, text generation device, computer equipment and computer readable storage medium
CN117521680A (en) * 2024-01-04 2024-02-06 深圳和成视讯科技有限公司 Online translation method of recorder and recorder

Similar Documents

Publication Publication Date Title
JP6818941B2 (en) How to Train Multilingual Speech Recognition Networks, Speech Recognition Systems and Multilingual Speech Recognition Systems
US20210390271A1 (en) Neural machine translation systems
US20240161732A1 (en) Multi-dialect and multilingual speech recognition
Tilk et al. Bidirectional Recurrent Neural Network with Attention Mechanism for Punctuation Restoration.
JP2020505650A (en) Voice recognition system and voice recognition method
CN113591462A (en) Bullet screen reply generation method and device and electronic equipment
Liu et al. Reinforcement learning for emotional text-to-speech synthesis with improved emotion discriminability
Khare et al. Low Resource ASR: The Surprising Effectiveness of High Resource Transliteration.
Nagaraj et al. Kannada to English Machine Translation Using Deep Neural Network.
KR20180062003A (en) Method of correcting speech recognition errors
CN112905772B (en) Semantic correlation analysis method and device and related products
Das et al. Best of both worlds: Robust accented speech recognition with adversarial transfer learning
CN112348073A (en) Polyphone recognition method and device, electronic equipment and storage medium
Granell et al. Multimodal crowdsourcing for transcribing handwritten documents
JP6810580B2 (en) Language model learning device and its program
CN116050425A (en) Method for establishing pre-training language model, text prediction method and device
Srinivasan et al. Multimodal speech recognition with unstructured audio masking
CN111508497B (en) Speech recognition method, device, electronic equipment and storage medium
Xie et al. Dialogue breakdown detection using hierarchical bi-directional LSTMs
CN111916052B (en) Voice synthesis method and device
KR102358485B1 (en) Dialogue system by automatic domain classfication
CN112989794A (en) Model training method and device, intelligent robot and storage medium
CN112686060B (en) Text translation method, device, electronic equipment and storage medium
KR20230156425A (en) Streaming ASR model delay reduction through self-alignment
Djeffal et al. Automatic speech recognition with BERT and CTC transformers: A review

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination