CN113220834A - Multimedia intelligent image matching method based on news content semantic analysis - Google Patents
Multimedia intelligent image matching method based on news content semantic analysis Download PDFInfo
- Publication number
- CN113220834A CN113220834A CN202110496811.5A CN202110496811A CN113220834A CN 113220834 A CN113220834 A CN 113220834A CN 202110496811 A CN202110496811 A CN 202110496811A CN 113220834 A CN113220834 A CN 113220834A
- Authority
- CN
- China
- Prior art keywords
- news
- matching
- picture
- matched
- pictures
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3346—Query execution using probabilistic model
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/50—Information retrieval; Database structures therefor; File system structures therefor of still image data
- G06F16/55—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/28—Quantising the image, e.g. histogram thresholding for discrimination between background and foreground patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/34—Smoothing or thinning of the pattern; Morphological operations; Skeletonisation
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02D—CLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
- Y02D10/00—Energy efficient computing, e.g. low power processors, power management or thermal management
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Databases & Information Systems (AREA)
- Multimedia (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Probability & Statistics with Applications (AREA)
- Computational Linguistics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses a multimedia intelligent image matching method based on news content semantic analysis, which comprises the following steps: s1 extracting the titles of the existing matching news and classifying; s2, classifying the pictures of the matched news according to the classification result of S1, and obtaining a first vector alpha; s3 reads the title and content of the unpatterned news and obtains a second vector beta; s4, determining the matching degree according to alpha and beta, word frequency, description and the like, and selecting alternative pictures according to the matching degree; s5, constructing a background picture library and a style picture library, accessing the alternative pictures into a semantic segmentation network, and processing to obtain background updated pictures; s6 performs style migration and smoothing on the background updated picture. The method can realize automatic matching, scene change and style migration of the news matching, improve news relevance of intelligent matching, save labor, expand the picture library, ensure the legality of picture sources and solve the problem of monotonous matching of the picture library to a certain extent.
Description
Technical Field
The invention belongs to the field of combination of natural language processing, information retrieval and image generation technologies, and particularly relates to a multimedia intelligent image matching method based on news content semantic analysis.
Background
In recent years, the combination of artificial intelligence technology and new internet media is more and more compact, and the news production and push efficiency is obviously improved. For example, in the field of news matching, most of traditional news matching needs to be edited and manually selected in a gallery, which is time-consuming and labor-consuming; the semantic feature description can be carried out on the pictures and news contents by a deep learning method, and automatic picture matching is realized by a semantic feature matching method, so that the picture matching efficiency is greatly improved. However, with the improvement of the protection consciousness of the copyright of the picture in the whole society, the news matching not only needs to consider the fitness of the picture and the news content, but also needs to improve the legal safety of the copyright of the used picture, which firstly requires the legality of the source of the picture library material, and simultaneously needs to consider the harmony of the whole image and text and the unity of style after matching. The existing matching method mainly focuses on the consistency of picture content and news content, and the legal safety problem of picture sources cannot be well solved. In addition, matching through a deep learning method also has the problems of low matching degree between pictures and news contents, insufficient coordination of styles and the like.
Disclosure of Invention
The invention aims to provide a multimedia intelligent matching method based on news content semantic analysis, which adopts the existing matching news pictures as materials to automatically classify and describe the pictures, realizes the automatic matching, scene change and style migration of the news matching pictures, improves the news relevance of the intelligent matching pictures, carries out the automatic matching of the media news without changing the original matching criteria, saves manpower, expands a picture library, ensures the legality of picture sources and solves the problem of monotonous matching picture library to a certain extent.
In order to solve the technical problems, the invention specifically adopts the following technical scheme:
a multimedia intelligent matching method based on news content semantic analysis comprises the following steps:
s1: extracting the titles of the existing matching picture news, and automatically classifying the titles by using a classifier;
s2: classifying the pictures of the existing matching news according to the classification result of S1, generating a first text description of the pictures of the existing matching news by the corresponding title, and then vectorizing the first text description to obtain a first vector alpha;
s3: reading the title and the content of the unpatterned news, classifying the title of the unpatterned news according to the classification result of S1, generating a second text description of the unpatterned news according to the title or the content of the unpatterned news, and then vectorizing the second text description to obtain a second vector beta;
s4: determining the matching degree of the first character description and the second character description according to the first vector alpha and the second vector beta and the word frequency of the first character description and the words with the same title or content as the un-matched news, and selecting a plurality of pictures with matched news as alternative pictures according to the matching degree;
s5: establishing a background picture library and a style picture library, accessing the alternative pictures into a semantic segmentation network, and performing background switching based on semantic segmentation by using a mask outputting a semantic segmentation result to obtain a background updating picture;
s6: and performing style migration and smoothing processing on the background updating picture.
Preferably, the classification in S1 is divided into two categories, one category is a fixed pattern matching diagram, and the other category is a variable pattern matching diagram.
Preferably, in S2, the first text description of the picture of the existing matching news is generated as follows: taking the title of the matched image news as the initial description of the image of the matched image news, and taking the union set of the titles of the matched image news as the initial description of the image of the matched image news for the titles of the matched image news with the same image of the matched image news; and according to the picture classification result, performing content extraction on the initial description of the picture of the matched picture news to generate a first text description of the picture of the matched picture news.
Preferably, in the S3, a second text description of the unpatterned news is generated as follows: if the fixed pattern matching is carried out, using a news title as an initial description of the un-matched news; if the mode is variable, using a textrank algorithm to extract the abstract of the unpatterned news, and simultaneously taking a union set with the headlines of the unpatterned news as an initial description of the unpatterned news; and then extracting the content of the initial description of the unpatterned news to generate a second text description of the unpatterned news.
Preferably, in S2, the first vector α is obtained by vectorizing the first text description using a pre-trained bert model, the second vector β is obtained by vectorizing the second text description using a pre-trained bert model in S3, and the cosine distance between the first vector α and the second vector β is calculated:
and then calculating the matching times of the same words of the first character description and the second character description:
matches(a,b)=match(a,b)/maxmatch
wherein a is a word extracted from the first text description, b is a word extracted from the second text description, matches (a, b) are the matching times of the same word in a, b, maxmatchIs the maximum value among all matches (a, b);
then, calculating the matching degree of the pictures of the matched picture news and the unmatched picture news:
l=matches(a,b)-k*dis(α,β);
wherein k is a system preset hyper-parameter;
and selecting a plurality of pictures in a preset interval of the maximum value in the matching degrees of all the pictures of the matched news and the unpatterned news as alternative pictures.
Preferably, in S5, the picture in the background library and the candidate picture are scaled to a uniform size by bilinear interpolation, and then the part with fixed background semantics is switched.
Preferably, training an automatic encoder taking image reconstruction as a loss function, sending the characteristics extracted by the automatic encoder into a WCT operator for transformation, and then decoding and restoring the characteristics into corresponding RGB space pictures; and defining the similar style of the similar pixels by using a gram matrix, then defining a target loss function by combining style difference with the similarity matrix of the pixels, and obtaining the picture with uniform style by optimizing the minimum target loss.
Preferably, the fixed pattern matching chart aims at the special column type news with fixed matching chart logic; the variable pattern thumbnails are directed to matching news that is related to both the headline and the news content.
Preferably, the content extraction is performed on the initial description of the image of the matched image news according to the following method, so as to generate a first text description of the image of the matched image news: if the pattern is matched in a fixed mode, performing rule matching on keywords by using a candidate word bank to generate the first text description; if the matched pictures are in the variable mode, extracting keywords by using TF-IDF (Transflash-inverse discrete frequency), wherein the TF range is the initial description of the pictures of the current matched pictures news, and the IDF range is the set of the initial descriptions of all the pictures of the matched pictures news
The invention has the following beneficial effects that firstly, the existing matching news is used as a material, the titles of the news are automatically classified, then the pictures are classified, and the first character description of the pictures is generated, for a news provider, the news provider has a large amount of matched pictures news, and the pictures are often manually edited, so that the image-text uniformity is good, the copyright source compliance is met, the picture description can be generated according to the news titles, and a basis is provided for later adaptation. The second text description is generated for the news title and the content without matching pictures, the matching degree of the picture and the news without matching pictures is calculated by the vectorization and the word frequency times of the first text description and the second text description, the alternative picture can be matched and matched with the content of the news without matching pictures, the picture and the news content can be more coordinated and consistent by performing background transformation and style migration on the picture, the number of pictures with different styles can be increased, and a material library is expanded. The method can be used for automatically matching the news, greatly saves labor and has good matching effect.
Detailed Description
For the purpose of facilitating an understanding of the present invention, the following detailed description is given with reference to the accompanying drawings and examples.
The multimedia intelligent image matching method based on news content analysis comprises the following steps:
s1, automatically classifying the titles of the existing matching news, which comprises the following steps:
s1-1, inputting a certain number of titles of the existing matching image news as a data set;
and S1-2, manually labeling the titles, and classifying the titles into 2 types, namely a fixed mode matching picture and a variable mode matching picture.
S1-3, using a fasttext algorithm to train and test the manually marked title to generate a classifier model;
s1-4, classifying all the headline data sets of the existing matching chart news by using the trained classifier model, namely classifying the headline data sets into two types of fixed mode matching charts and variable mode matching charts;
s2, establishing an image classification library and a corresponding picture description, and specifically comprising the following steps:
s2-1, classifying the pictures of the existing matching image news according to the classification result of the existing matching image news headlines in S1, namely classifying the pictures into two classes of a fixed mode matching image and a variable mode matching image;
s2-2, taking the title of the existing matching picture news as the initial description of the corresponding picture (original matching picture), and taking the union of the titles as the initial description for the title with the same original matching picture;
and S2-3, extracting the content of the initial description according to the picture classification result of the existing matching news to generate a first character description (namely a final picture description) of the picture.
S3, analyzing the news content without matching pictures to generate a key content description, namely a second text description, and the specific steps are as follows:
s3-1, reading the content and title of the news article without the matching picture;
s3-2, classifying the undepicted news headlines, wherein if the images are matched in a fixed mode, the undepicted news headlines are used as initial descriptions of the undepicted news articles; if the variable mode map is matched, abstract extraction is carried out on the content of the non-matched map news article by using a textrank algorithm, and meanwhile, a union set of the abstract extraction and the title of the non-matched map news article is taken as an initial description of the non-matched map news article;
s3-3, extracting the content of the initial description of the unpatterned news article to generate a second text description (namely the final article description of the unpatterned news);
s4, matching the images through text matching search, which comprises the following steps:
s4-1, vectorizing a text description (namely a first character description) corresponding to a picture database of the existing matching news by using a pre-training bert model to obtain a first vector alpha (in the step, the first character description can be obtained at S2-3 and then vectorized);
and S4-2, vectorizing the second word description extracted in S3 by using a bert model pre-trained by the same corpus as that in S4-1 to obtain a second vector beta (generating the second vector beta can also be carried out after obtaining the second word description in S3-3). The cosine distance and the word frequency matching jointly define the matching response degree of the picture text description and the news content, and the cosine distance defines: as shown in the following formula, where α and β represent a text vector of a news keyword and a text description vector of a picture, respectively, which are vectorized by the bert model.
Word frequency matching definition: a and b are respectively sentences described by the picture texts and words extracted from news contents, and match is defined as the number of times of matching of words in the a and b. maxmatchDefined as the maximum value of match in all data.
matches(a,b)=match(a,b)/maxmatch
The matching degree of a graph and news content can be defined as follows: where k is a hyperparameter to balance the contributions of the two different distance definitions, the system may preset k, e.g. to 0.7.
l=matches(a,b)-k*dis(α,β)
By calculating the matching degree of each picture in the picture data set with the news content, the largest 10 pictures can be selected as a candidate.
S5, switching picture backgrounds, which comprises the following steps:
s5-1, a background picture library and a style picture library are constructed. And constructing a background picture and a style picture library according to the requirement and the news content.
And S5-2, accessing the alternative picture obtained in the step S4 into a semantic segmentation network, and performing background switching based on semantic segmentation by using a mask for outputting a semantic segmentation result. Scaling the background picture and the alternative picture to a uniform size in a bilinear interpolation mode, and then switching the part with fixed background semantics. For example, a background switching of the sky may be made for outdoor scenes. That is, the picture portion of the sky in the mask is switched to the pixels corresponding to the background picture of the sky.
S6, transferring the style of the picture, which comprises the following steps:
s6-1 style migration: the method comprises the steps of training an automatic encoder with image reconstruction as a loss function, sending features extracted by the encoder into a WCT operator for transformation, and then decoding and restoring the features into corresponding RGB space pictures. And obtaining the picture with small style difference by optimizing the image reconstruction loss function to be smaller.
S6-2 image smoothing: the picture with the uniform style can be defined as that similar pixels in a local area have the similar style, a gram matrix is used for defining the similar style of the similar pixels, then a target loss function can be defined by combining style difference with the similarity matrix of the pixels, and the picture which is consistent with the uniform style is obtained by optimizing the minimum target loss. We define where yiRepresenting the pixel output of the previous step in the optimization process, riRepresenting the pixels in the map R resulting from this step in the optimization process. The definitions of W and d are as follows.
W={wij}∈Rn·n
The invention selects pictures from the existing matching news as the matching materials, solves the problem of picture copyright, further realizes automatic matching, increases the diversity of the pictures to a certain extent through background conversion and style migration, and expands the picture set. The method can be applied to the scene of media news matching demand, saves labor and realizes an objective evaluation matching mode.
The above embodiments are merely illustrative of the concept and implementation of the present invention, and are not restrictive, and technical solutions that are not substantially changed under the concept of the present invention are still within the scope of protection.
Claims (9)
1. A multimedia intelligent matching method based on news content semantic analysis is characterized by comprising the following steps:
s1: extracting the titles of the existing matching picture news, and automatically classifying the titles by using a classifier;
s2: classifying the pictures of the existing matching news according to the classification result of S1, generating a first text description of the pictures of the existing matching news by the corresponding title, and then vectorizing the first text description to obtain a first vector alpha;
s3: reading the title and the content of the unpatterned news, classifying the title of the unpatterned news according to the classification result of S1, generating a second text description of the unpatterned news according to the title or the content of the unpatterned news, and then vectorizing the second text description to obtain a second vector beta;
s4: determining the matching degree of the first character description and the second character description according to the first vector alpha and the second vector beta and the word frequency of the first character description and the words with the same title or content as the un-matched news, and selecting a plurality of pictures with matched news as alternative pictures according to the matching degree;
s5: establishing a background picture library and a style picture library, accessing the alternative pictures into a semantic segmentation network, and performing background switching based on semantic segmentation by using a mask outputting a semantic segmentation result to obtain a background updating picture;
s6: and performing style migration and smoothing processing on the background updating picture.
2. The method as claimed in claim 1, wherein the classification in S1 is divided into two categories, one category is fixed pattern matching and the other category is variable pattern matching.
3. The method for multimedia intelligent matching of images based on semantic analysis of news contents as claimed in claim 1, wherein in S2, the first text description of the image of the existing matching news is generated as follows: taking the title of the matched image news as the initial description of the image of the matched image news, and taking the union set of the titles of the matched image news as the initial description of the image of the matched image news for the titles of the matched image news with the same image of the matched image news; and according to the picture classification result, performing content extraction on the initial description of the picture of the matched picture news to generate a first text description of the picture of the matched picture news.
4. The method for multimedia intelligent matching of images based on semantic analysis of news contents according to claim 2, wherein in the step S3, a second text description of unpatterned news is generated as follows: if the fixed pattern matching is carried out, using a news title as an initial description of the un-matched news; if the mode is variable, using a textrank algorithm to extract the abstract of the unpatterned news, and simultaneously taking a union set with the headlines of the unpatterned news as an initial description of the unpatterned news; and then extracting the content of the initial description of the unpatterned news to generate a second text description of the unpatterned news.
5. The method as claimed in claim 1, wherein in S2, the first vector α is obtained by vectorizing the first text description using a pre-trained bert model, in S3, the second vector β is obtained by vectorizing the second text description using a pre-trained bert model, and cosine distances of the first vector α and the second vector β are calculated:
and then calculating the matching times of the same words of the first character description and the second character description:
matches(a,b)=match(a,b)/maxmatch
wherein a is a word extracted from the first text description, b is a word extracted from the second text description, matches (a, b) are the matching times of the same word in a, b, maxmatchIs the maximum value among all matches (a, b);
then, calculating the matching degree of the pictures of the matched picture news and the unmatched picture news:
l=matches(a,b)-k*dis(α,β);
wherein k is a system preset hyper-parameter;
and selecting a plurality of pictures in a preset interval of the maximum value in the matching degrees of all the pictures of the matched news and the unpatterned news as alternative pictures.
6. The method as claimed in claim 1, wherein in S5, the pictures in the background library and the candidate pictures are scaled to a uniform size by bilinear interpolation, and then the parts with fixed background semantics are switched.
7. The intelligent multimedia mapping method based on semantic analysis of news contents as claimed in claim 1, wherein the style migration and smoothing process is performed as follows:
training an automatic encoder taking image reconstruction as a loss function, sending the characteristics extracted by the automatic encoder into a WCT operator for transformation, and then decoding and restoring the characteristics into corresponding RGB space pictures; and defining the similar style of the similar pixels by using a gram matrix, then defining a target loss function by combining style difference with the similarity matrix of the pixels, and obtaining the picture with uniform style by optimizing the minimum target loss.
8. The multimedia intelligent matching method based on news content semantic analysis as claimed in claim 2, wherein the fixed pattern matching is for special column type news with fixed matching logic; the variable pattern thumbnails are directed to matching news that is related to both the headline and the news content.
9. The multimedia intelligent matching method based on news content semantic analysis as claimed in claim 2, wherein content extraction is performed on the initial description of the picture of the matched picture news according to the following method to generate a first text description of the picture of the matched picture news: if the pattern is matched in a fixed mode, performing rule matching on keywords by using a candidate word bank to generate the first text description; and if the image is matched in the variable mode, extracting keywords by using TF-IDF, wherein the TF range is the initial description of the image of the current matched image news, and the IDF range is the set of the initial descriptions of all the images of the matched image news.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110496811.5A CN113220834B (en) | 2021-05-07 | 2021-05-07 | Multimedia intelligent picture allocation method based on news content semantic analysis |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110496811.5A CN113220834B (en) | 2021-05-07 | 2021-05-07 | Multimedia intelligent picture allocation method based on news content semantic analysis |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113220834A true CN113220834A (en) | 2021-08-06 |
CN113220834B CN113220834B (en) | 2023-08-29 |
Family
ID=77091678
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110496811.5A Active CN113220834B (en) | 2021-05-07 | 2021-05-07 | Multimedia intelligent picture allocation method based on news content semantic analysis |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113220834B (en) |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090322943A1 (en) * | 2008-06-30 | 2009-12-31 | Kabushiki Kaisha Toshiba | Telop collecting apparatus and telop collecting method |
CN110136226A (en) * | 2019-04-08 | 2019-08-16 | 华南理工大学 | It is a kind of to cooperate with the news of description generation to match drawing method automatically based on image group |
CN112241481A (en) * | 2020-10-09 | 2021-01-19 | 中国人民解放军国防科技大学 | Cross-modal news event classification method and system based on graph neural network |
-
2021
- 2021-05-07 CN CN202110496811.5A patent/CN113220834B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090322943A1 (en) * | 2008-06-30 | 2009-12-31 | Kabushiki Kaisha Toshiba | Telop collecting apparatus and telop collecting method |
CN110136226A (en) * | 2019-04-08 | 2019-08-16 | 华南理工大学 | It is a kind of to cooperate with the news of description generation to match drawing method automatically based on image group |
CN112241481A (en) * | 2020-10-09 | 2021-01-19 | 中国人民解放军国防科技大学 | Cross-modal news event classification method and system based on graph neural network |
Non-Patent Citations (1)
Title |
---|
任海平;: "深度学习下智能写稿辅助服务的设计与实践", 传媒论坛, no. 04 * |
Also Published As
Publication number | Publication date |
---|---|
CN113220834B (en) | 2023-08-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111858954B (en) | Task-oriented text-generated image network model | |
CN101573705B (en) | Media material analysis of continuing article portions | |
CN110275964A (en) | The recommended models of knowledge based map and Recognition with Recurrent Neural Network | |
CN109271537B (en) | Text-to-image generation method and system based on distillation learning | |
WO2023065617A1 (en) | Cross-modal retrieval system and method based on pre-training model and recall and ranking | |
CN108984642B (en) | Printed fabric image retrieval method based on Hash coding | |
Li et al. | Residual attention-based LSTM for video captioning | |
CN107862561A (en) | A kind of method and apparatus that user-interest library is established based on picture attribute extraction | |
CN114996488A (en) | Skynet big data decision-level fusion method | |
CN112633431B (en) | Tibetan-Chinese bilingual scene character recognition method based on CRNN and CTC | |
CN112541347B (en) | Machine reading understanding method based on pre-training model | |
CN116975615A (en) | Task prediction method and device based on video multi-mode information | |
CN116010552A (en) | Engineering cost data analysis system and method based on keyword word library | |
Fei et al. | Learning user interest with improved triplet deep ranking and web-image priors for topic-related video summarization | |
CN117556067A (en) | Data retrieval method, device, computer equipment and storage medium | |
CN113254688A (en) | Trademark retrieval method based on deep hash | |
Liu et al. | Bi-lstm sequence modeling for on-the-fly fine-grained sketch-based image retrieval | |
CN113220834B (en) | Multimedia intelligent picture allocation method based on news content semantic analysis | |
KR102586580B1 (en) | News editing supoort system using natural language processing artificial intelligence language model | |
Cheu | Construction of a digital platform for Nuo opera based on artificial intelligence | |
Liu et al. | Entity representation learning with multimodal neighbors for link prediction in knowledge graph | |
Pang et al. | Chinese Calligraphy Character Image Recognition and Its Applications in Web and Wechat Applet Platform | |
CN109062995B (en) | Personalized recommendation algorithm for drawing Board (Board) cover on social strategy exhibition network | |
Jia et al. | Printed score detection based on deep learning | |
Aigrain et al. | A connection graph for user navigation in a large image bank. |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |