CN103714178B - Automatic image marking method based on word correlation - Google Patents

Automatic image marking method based on word correlation Download PDF

Info

Publication number
CN103714178B
CN103714178B CN201410008553.1A CN201410008553A CN103714178B CN 103714178 B CN103714178 B CN 103714178B CN 201410008553 A CN201410008553 A CN 201410008553A CN 103714178 B CN103714178 B CN 103714178B
Authority
CN
China
Prior art keywords
image
word
words
training set
annotation
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201410008553.1A
Other languages
Chinese (zh)
Other versions
CN103714178A (en
Inventor
安震
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Jingdong Century Trading Co Ltd
Beijing Jingdong Shangke Information Technology Co Ltd
Original Assignee
Beijing Jingdong Century Trading Co Ltd
Beijing Jingdong Shangke Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Jingdong Century Trading Co Ltd, Beijing Jingdong Shangke Information Technology Co Ltd filed Critical Beijing Jingdong Century Trading Co Ltd
Priority to CN201410008553.1A priority Critical patent/CN103714178B/en
Publication of CN103714178A publication Critical patent/CN103714178A/en
Application granted granted Critical
Publication of CN103714178B publication Critical patent/CN103714178B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2155Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Databases & Information Systems (AREA)
  • Library & Information Science (AREA)
  • Artificial Intelligence (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an automatic image marking method based on work correlation. A training set T comprises l images, n marking words are marked on each image of the training set T, the training set T is provided with a corresponding vision lemma, and the image to be marked is I. The method includes the steps that a semantic vector of each marking word w is calculated according to a formula, the marking word w is represented by the vector form w=<v1, v2,......, vm>, ci is an associated word in a context, and the associated words in the context total m; semantic similarity of the marking words is calculated according to a formula, and vector module calculated is achieved as is shown in the specification; p(A) is calculated according to the formula, wherein A is a marking word group in w1, w2,......wn, and n is the number of the marking word groups; the conditional probability p(I/wi) is calculated according to a formula; the marking word group A of the image I to be marked is calculated according to the formula A=arg maxAp(I/A) p(A).

Description

Automatic image labeling method based on inter-word correlation
Technical Field
The invention relates to the field of image processing, in particular to an automatic image annotation method based on interword correlation.
Background
With the rapid development of multimedia and internet technologies, people's daily life and work have increasingly strong dependence on multimedia information such as images. The semantic-based image retrieval not only can accurately express the retrieval intention of the user, but also is convenient for the user to use, so that the retrieval mode not only becomes an important form of image retrieval, but also becomes a technical hotspot pursued by researchers.
The image automatic labeling technology is an important and challenging work in image semantic retrieval, and the image automatic labeling technology is presented for automatically acquiring semantic information contained in image visual content, and tries to build a bridge between image bottom-layer visual features and high-layer semantics, so as to support semantic retrieval at a semantic level. Therefore, the automatic annotation algorithm research based on the image semantics becomes a very active research branch and key technology in the image retrieval field, and has good application prospect and research value.
The automatic image annotation means that a computer automatically adds semantic keywords capable of reflecting the image content to an unmarked image. The method automatically practices a relation model of a semantic concept space and a visual characteristic space by using a labeled image set or other available information, and labels an image with unknown semantics by using the model. The semantic gap problem is solved to a certain extent by establishing a mapping relation between the high-level semantic information and the bottom-level features of the image.
The image automatic labeling method of the joint media related model is an image labeling algorithm which is most widely applied in the image labeling method based on the generated model at present, and is widely researched by scholars. The basic idea of the annotation model is to establish probability correlation between an image visual characteristic space and a semantic concept space by using a probability statistics method, find out a group of semantic annotation words to enable the joint probability between the semantic annotation words and image contents to be maximum by statistically learning joint probability distribution between the image visual characteristic space and the semantic concept space, and use the group of annotation words as the final annotation of an image to be detected.
However, the joint media correlation model belongs to a probability model, and the model has a bias on the tagging words with high occurrence frequency. Secondly, in the automatic annotation method of the joint media related model image, different candidate annotation words are assumed to be mutually independent in the annotation process, and the relevance between the annotation words is not fully mined. In fact, in the same image, there are various associations such as symbiosis, hierarchy or space between different annotation words.
For example, an image containing semantic objects such as "sun," sky, crowd, mountain, tree "and the like, it can be seen from the visual content of the image that there is a certain spatial correlation between the" sun "and" sky "objects, and" sun "cannot exist independently from the semantic object of" sky "; similarly, for two semantic objects of "mountain" and "tree" in the image content, the "tree" object exists with the "mountain" semantic object as the visual content background, and the two semantic objects are also in inseparable relation in the image visual content, and it cannot be absolutely assumed that the two annotation words are annotated independently from each other. Therefore, the automatic annotation algorithm of the joint media related model image considers that the method of mutually independence among different candidate annotation words has certain defects in the annotation process, and the phenomenon of inconsistent semantics among the annotation words in the annotation result can be caused by neglecting the correlation among the words.
Disclosure of Invention
In view of the above, the present invention provides an automatic image annotation method based on inter-word correlation, so as to overcome the defect of the joint media correlation model image automatic annotation algorithm in the annotation process that different candidate annotation words are considered to be independent from each other, and solve the problem of semantic inconsistency between the annotation words in the annotation result due to neglect of inter-word correlation. The technical scheme provided by the invention is as follows:
an image automatic labeling method based on word correlation is disclosed, wherein a training set T comprises l images, and the l images form an image set P ═ P1p2… pl](ii) a Each image of the training set T is marked with n tagging words, and all the tagging words in the training set T form a tagging word set W ═ W1w2… ws](ii) a Each image in the training set T has a corresponding visual lemma, and all the visual lemmas in the training set T form a visual lemma set B ═ B1b2… by]And the image to be marked is I, and the method comprises the following steps:
A. according to the formulaCalculating a semantic vector of each annotation word w in the training set T, and expressing the annotation word w as a vector form w ═ w<v1,v2,…,vm>Wherein c isiFor context-related words, there are m context-related words, p (c)i) For context associated words ciOverall distribution probability of p (c)i/w) represents a context-related word ciAnd the ratio of the number of co-occurrences of the annotation word w in the training set T to the total number of occurrences of the annotation word w in the training set T, i.e.The context associated word is a label word in a training set T;
B. according to the formulaCalculating semantic similarity between the labeled words, wherein | | · | | is calculated by a vector model, and wi·wjCalculating vector dot product;
C. according to the formulaCalculating p (A), wherein A is a labeling phrase { w1,w2,…wnN is the number of the labeled phrases;
D. according to the formulaCalculating the conditional probability p (I/w)i) Wherein p (w)i) To mark a word wiThe ratio of the number of occurrences in the training set T to the total number of occurrences of all the tokens in the training set T, i.e. the ratio
p(wi,b1,…,bn) The calculation method comprises the following steps:
wherein P (J) represents the probability of randomly drawing a training image J in the image set P; p (w)i/J) representing the occurrence of words w in the training image JiA posterior probability of (d); and p (b)k/J) represents a training image JIn the middle of which appears visual word element bkA posterior probability of (d);
E. according toCalculating p (I/A);
F. from formula A ═ arg maxAp (I/A) p (A) calculating an annotation phrase A of the image I to be annotated.
In the above scheme, p (w) in step Di/J) and p (b)kThe calculation methods of/J) are respectively as follows:
p ( w i | J ) = ( 1 - &alpha; J ) # ( w i , J ) | J | + &alpha; J # ( w i , T ) | T | - - - ( 1 )
p ( b k | J ) = ( 1 - &beta; J ) # ( b k , J ) | J | + &beta; J # ( b k , T ) | T | - - - ( 2 )
wherein, αJAnd βJIs an empirical value for the smoothing parameter;
#(wiand J) represents a notation word wiWhether it appears in training image J, and if so, # wiJ) 1, otherwise # (w)i,J)=0;
#(wiT) denotes a label word wiIf it is present in the training set T, then # (w)iT) 1, otherwise # (w)i,T)=0;
#(bkAnd J) visual morpheme bkWhether it appears in training image J, and if so # (b)kJ) 1, otherwise # (b)k,J)=0;
The | J | represents the total number of the labeled words and the visual word elements in the training image J; and | T | represents the total number of the labeled words and the visual word elements in the training set T.
In summary, the technical scheme provided by the invention converts the joint probability calculation process of the annotation words and the images in the joint media correlation model into the solution of the probability of the occurrence of the images and the prior probability of the annotation word groups under the condition of the annotation words, thereby greatly reducing the influence of the high-frequency candidate annotation words on the probability statistic model, enabling the non-high-frequency candidate annotation words to play a greater role, improving the recall ratio and precision ratio of the non-high-frequency candidate annotation words, simultaneously introducing the semantic similar language model into the joint media correlation model, and estimating the prior probability of a group of annotation words through the semantic similar language model, so that a group of annotation words with stronger semantic correlation are more likely to be generated. Thereby improving the overall labeling effect of the image.
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention will be further described in detail with reference to the accompanying drawings and specific embodiments.
The technical scheme of the invention is as follows:
A. according to the formulaCalculating a semantic vector of each annotation word w in the training set T, and expressing the annotation word w as a vector form w ═ w<v1,v2,…,vm>Wherein c isiFor the context-related words, there are m context-related words in total,p(ci) For context associated words ciOverall distribution probability of p (c)i/w) represents a context-related word ciAnd the ratio of the number of co-occurrences of the annotation word w in the training set T to the total number of occurrences of the annotation word w in the training set T, i.e.The context associated word is a label word in a training set T;
B. according to the formulaCalculating semantic similarity between the labeled words, wherein | | · | | is calculated by a vector model, and wi·wjCalculating vector dot product;
C. according to the formulaCalculating p (A), wherein A is a labeling phrase { w1,w2,…wnN is the number of the labeled phrases;
D. according to the formulaCalculating the conditional probability p (I/w)i) Wherein p (w)i) To mark a word wiThe ratio of the number of occurrences in the training set T to the total number of occurrences of all the tokens in the training set T, i.e. the ratio
p(wi,b1,…,bn) The calculation method comprises the following steps:
wherein P (J) represents the probability of randomly drawing a training image J in the image set P; p (w)i/J) represents the training image JThe existing vocabulary wiA posterior probability of (d); and p (b)k/J) visual lemma b appearing in training image JkA posterior probability of (d);
E. according toCalculating p (I/A);
F. from formula A ═ arg maxAp (I/A) p (A) calculating an annotation phrase A of the image I to be annotated.
The current image annotation problem can be defined as: giving a training set T, wherein the training set T comprises an image set P and a tagging word set W, and each image PiHow to select a group of annotation words a from the annotation word set W to label a new image I?
The image labeling method of the invention adopts a probability model, the target is to find a labeling phrase A, the conditional probability p (A/I) is maximum, namely:
A=arg maxAp(A/I) (3)
wherein A is a markup phrase { w }1,w2,…wnImage I with a set of visual features b1,b2…,bmThe representation is obtained by performing preprocessing (such as image segmentation, feature extraction, feature value normalization and other operations) and image block region classification operation on the image I. p (A/I) can be rewritten as follows:
p ( A / I ) = p ( A , I ) p ( I ) - - - ( 4 )
since the prior probability of an image is usually considered to be uniformly distributed, p (I) can be considered as a constant, and
p(A,I)=p(I/A)p(A) (5)
the formula (3) is simplified by the formulas (4) and (5), and the following results are obtained:
A=arg maxAp(I/A)p(A) (6)
and solving the maximum value by combining the two probabilities of p (I/A) and p (A) to find the optimal annotation phrase A. p (I/A) can be derived from the original image annotation model, and p (A) can be derived from the language model. The influence of the original image model and the language model on the finally obtained labeling effect is represented by giving different weights to the two probabilities:
A = arg max A p ( I / A ) &lambda; 1 p ( A ) &lambda; 2 - - - ( 7 )
it is converted into the following form:
A=arg maxA1log p(I/A)+λ2log p(A)) (8)
the annotation phrase A can be obtained by calculating p (A) and p (I/A). Wherein λ is1And λ2The method is determined in the machine learning and model building process of a training image set, and two constants are used in the automatic labeling process of the image to be detected.
The technical solution of the present invention is described below by taking a training set T containing l images as an example, where an image to be labeled is I. The training set T has one image forming set P ═ P1p2… pl](ii) a Each image of the training set T is marked with n tagging words, and all the tagging words in the training set T form a tagging word set W ═ W1w2… ws](ii) a Each image in the training set T has a corresponding visual lemma, and all the visual lemmas in the training set T form a visual lemma set B ═ B1b2… by]。
Fig. 1 is a flowchart of the present embodiment, and as shown in fig. 1, the method includes the following steps:
step 101: and carrying out image preprocessing and block region classification operation on the image I to be annotated.
In the step, image preprocessing (image segmentation, feature extraction, feature value normalization and the like) is carried out on an image I to be annotated, then image block region classification operation is carried out, a clustering algorithm is utilized to classify each image block region, and visual word element combination is used for representing image visual content: i ═ I1i2… if}. The method for obtaining visual lemmas is prior art and will not be described in detail here.
Step 102: p (a) is calculated by a semantic similar language model.
In order to introduce the relevance information among the annotation words into the similarity among the annotation words, the invention adopts a semantic vector model to represent each annotation word w: context-related word set C ═ C1c2… cm]Each element ofciThe method represents one context related word, m context related words are provided in total, and all the annotation words in the annotation word set W in the training set T can be selected as the context related words, that is, C is W. Each annotation word w is represented by a context-related word vector associated therewith, i.e. w ═<v1,v2,…,vm>Wherein the calculation of each semantic component vi is defined as a context-related word ciConditional probability and context associated word c relative to annotation word wiRatio of probabilities of (c):
v i = p ( c i / w ) p ( c i ) - - - ( 9 )
wherein p (c)i) Representing a context-related word ciThe overall distribution probability of (2) is uniform distribution. Conditional probability p (c)i/w) represents a context-related word ciAnd the ratio of the co-occurrence times of the annotation word w when all the images in the image set P in the training set T are labeled to the total times of the annotation word w when all the images in the image set P are labeled:
p ( c i / w ) = c o u n t ( c i , w ) c o u n t ( w ) - - - ( 10 )
p(ci/w) represents the intensity distribution of the co-occurrence of the word w and the context-related word, and the division by the overall probability of each context-related word is to prevent the semantic vector w from being formed<v1,v2,…,vm>Is dominated by context relevant words with high frequency of occurrence, because the relevant words with high frequency often have large conditional probability. As shown in table 1, where "sky", "sun", "clouds" and "town" represent a group of context-related words, "tree", "building" and "river" are a group of annotation words, and the semantic vectors of the annotation words are shown in table 1.
TABLE 1
sky sun clouds town
tree 2.56 0.91 0.74 0.63
building 5.01 0.57 2.41 21.19
river 2.57 2.57 1.12 5.72
Semantic similarity between the annotated words is then calculated. The similarity is calculated as shown in equation 11:
s i m ( w i , w j ) = w i &CenterDot; w j | | w i | | &CenterDot; | | w j | | - - - ( 11 )
where | | · | | is calculated modulo the vector.
wi·wjThe calculation of (d) is shown in equation 12:
w i &CenterDot; w j = &Sigma; k = 1 m v w i , k v w j , k = &Sigma; k = 1 m p ( c k / w i ) p ( c k ) &CenterDot; p ( c k / w j ) p ( c k ) - - - ( 12 )
wherein c iskRepresenting a context-associated word. Semantic similarity between the annotated words is shown in table 2. The similarity value range is 0 to 1, and the higher the numerical value is, the higher the similarity between two annotation words is, the higher the probability that the two annotation words appear in the same image is.
TABLE 2
Assuming that in the same annotation the annotation word is semantically related to the context-related word, a set of annotation words a ═ w1,w2,…,wnThe probability p (a) of each annotation word can be obtained by calculating the similarity between each annotation word and other annotation words:
p ( A ) &Proportional; 1 n - 1 &Sigma; w i &Element; A &Sigma; w j &Element; A , j &NotEqual; i s i m ( w i , w j ) - - - ( 13 )
by substituting equations 10, 11, and 12 into equation 13, the probability p (a) of the labeled phrase can be calculated:
p ( A ) &Proportional; 1 n - 1 &Sigma; w i &Element; A &Sigma; w j &Element; A , j &NotEqual; i &Sigma; k = 1 m c o u n t ( c k , w i ) c o u n t ( w i ) &CenterDot; p ( c k ) &CenterDot; c o u n t ( c k , w j ) c o u n t ( w j ) &CenterDot; p ( c k ) | | w i | | &CenterDot; | | w j | | - - - ( 14 )
step 103: p (I/A) is calculated by the joint media correlation model.
In this step, firstly, the formula is based onCalculating the conditional probability p (I/w)i). Wherein,
p(wi) The calculation method comprises the following steps:
by marking words wiThe ratio of the number of occurrences in the training set T to the total number of occurrences of all the annotation words represents the vocabulary wiPrior probability p (w)i):
p ( w i ) = | w i | &Sigma; w k &Element; T | w k | - - - ( 15 )
p(wi,b1,…,bn) The calculation method comprises the following steps:
p ( w i , b 1 , b 2 , ... , b n ) = &Sigma; J &Element; T p ( J ) p ( w i | J ) &Pi; k = 1 n p ( b k | J ) - - - ( 16 )
p (J) represents the probability of randomly drawing a training image J in the image set P, which is generally assumed to be uniformly distributed; p (w)i/J) representing the occurrence of words w in the training image JiA posterior probability of (d); whilep(bk/J) visual lemma b appearing in training image JkThe posterior probability of (d). The probability value for each term is estimated as follows:
p ( w i | J ) = ( 1 - &alpha; J ) # ( w i , J ) | J | + &alpha; J # ( w i , T ) | T | - - - ( 17 )
p ( b k | J ) = ( 1 - &beta; J ) # ( b k , J ) | J | + &beta; J # ( b k , T ) | T | - - - ( 18 )
wherein, αJAnd βJIs an empirical value for the smoothing parameter; # (w)iAnd J) represents a notation word wiWhether it appears in training image J, and if so, # wiJ) 1, otherwise # (w)i,J)=0;#(wiT) denotes a label word wiIf it is present in the training set T, then # (w)iT) 1, otherwise # (w)i,T)=0;#(bkAnd J) visual morpheme bkWhether it appears in training image J, and if so # (b)kJ) 1, otherwise # (b)kJ) is 0; the | J | represents the total number of the labeled words and the visual word elements in the training image J; and | T | represents the total number of the labeled words and the visual word elements in the training set T.
Then p (I/A) can be approximated as
Step 104: and calculating the phrase to be labeled of the image I to be labeled.
Above, p (A) and p (I/A) are solved, respectively, according to A ═ arg maxA1log p(I/A)+λ2log p (A)) can calculate the annotation phrase A for an image I
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that are within the spirit and principle of the present invention are intended to be included within the scope of the present invention.

Claims (2)

1. An image automatic labeling method based on the correlation between words is characterized in that a training set T comprises l images, and the l images form an image set P ═ P1p2… pl](ii) a Each image of the training set T is marked with n tagging words, and all the tagging words in the training set T form a tagging word set W ═ W1w2… ws](ii) a Each image in the training set T has a corresponding visual lemma, and all the visual lemmas in the training set T form a visual lemma set B ═ B1b2… by]The image to be labeled isI, the method comprises the following steps:
A. according to the formulaCalculating a semantic vector of each annotation word w in the training set T, and expressing the annotation word w as a vector form w ═ w<v1,v2,…,vm>Wherein c isiFor context-related words, there are m context-related words, p (c)i) For context associated words ciOverall distribution probability of p (c)i/w) represents a context-related word ciAnd the ratio of the number of co-occurrences of the annotation word w in the training set T to the total number of occurrences of the annotation word w in the training set T, i.e.The context associated word is a label word in a training set T;
B. according to the formulaCalculating semantic similarity between the labeled words, wherein | | · | | is calculated by a vector model, and wi·wjCalculating vector dot product;
C. according to the formulaCalculating p (A), wherein A is a labeling phrase { w1,w2,…wnN is the number of the labeled phrases;
D. according to the formulaCalculating the conditional probability p (I/w)i) Wherein p (w)i) To mark a word wiThe ratio of the number of occurrences in the training set T to the total number of occurrences of all the tokens in the training set T, i.e. the ratio
p(wi,b1,…,bn) The calculation method comprises the following steps:
wherein P (J) represents the probability of randomly drawing a training image J in the image set P; p (w)i/J) representing the occurrence of words w in the training image JiA posterior probability of (d); and p (b)k/J) visual lemma b appearing in training image JkA posterior probability of (d);
E. according toCalculating p (I/A);
F. from formula A ═ argmaxAp (I/A) p (A) calculating an annotation phrase A of the image I to be annotated.
2. The method of claim 1, wherein p (w) in step Di/J) and p (b)kThe calculation methods of/J) are respectively as follows:
p ( w i | J ) = ( 1 - &alpha; J ) # ( w i , J ) | J | + &alpha; J # ( w i , T ) | T | - - - ( 1 )
p ( b k | J ) = ( 1 - &beta; J ) # ( b k , J ) | J | + &beta; J # ( b k , T ) | T | - - - ( 2 )
wherein, αJAnd βJIs an empirical value for the smoothing parameter;
#(wiand J) represents a notation word wiWhether it appears in training image J, and if so, # wiJ) 1, otherwise # (w)i,J)=0;
#(wiT) denotes a label word wiIf it is present in the training set T, then # (w)iT) 1, otherwise # (w)i,T)=0;
#(bkAnd J) visual morpheme bkWhether it appears in training image J, and if so # (b)kJ) 1, otherwise # (b)k,J)=0;
The | J | represents the total number of the labeled words and the visual word elements in the training image J; and | T | represents the total number of the labeled words and the visual word elements in the training set T.
CN201410008553.1A 2014-01-08 2014-01-08 Automatic image marking method based on word correlation Active CN103714178B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410008553.1A CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410008553.1A CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Publications (2)

Publication Number Publication Date
CN103714178A CN103714178A (en) 2014-04-09
CN103714178B true CN103714178B (en) 2017-01-25

Family

ID=50407153

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410008553.1A Active CN103714178B (en) 2014-01-08 2014-01-08 Automatic image marking method based on word correlation

Country Status (1)

Country Link
CN (1) CN103714178B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104794183A (en) * 2015-04-10 2015-07-22 浙江大学 Picture labeling method based on multiple views and multiple labels
CN108604902A (en) * 2016-02-08 2018-09-28 皇家飞利浦有限公司 Determine the device and method of cluster
CN108268875B (en) * 2016-12-30 2020-12-08 广东精点数据科技股份有限公司 Image semantic automatic labeling method and device based on data smoothing
CN110162644B (en) * 2018-10-10 2022-12-20 腾讯科技(深圳)有限公司 Image set establishing method, device and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920820A (en) * 2006-09-14 2007-02-28 浙江大学 Image meaning automatic marking method based on marking significance sequence
CN101620615A (en) * 2009-08-04 2010-01-06 西南交通大学 Automatic image annotation and translation method based on decision tree learning
CN101685464A (en) * 2009-06-18 2010-03-31 浙江大学 Method for automatically labeling images based on community potential subject excavation
CN102298606A (en) * 2011-06-01 2011-12-28 清华大学 Random walking image automatic annotation method and device based on label graph model
CN102542067A (en) * 2012-01-06 2012-07-04 上海交通大学 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101587478B (en) * 2008-05-20 2013-07-24 株式会社理光 Methods and devices for training, automatically labeling and searching images

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1920820A (en) * 2006-09-14 2007-02-28 浙江大学 Image meaning automatic marking method based on marking significance sequence
CN101685464A (en) * 2009-06-18 2010-03-31 浙江大学 Method for automatically labeling images based on community potential subject excavation
CN101620615A (en) * 2009-08-04 2010-01-06 西南交通大学 Automatic image annotation and translation method based on decision tree learning
CN102298606A (en) * 2011-06-01 2011-12-28 清华大学 Random walking image automatic annotation method and device based on label graph model
CN102542067A (en) * 2012-01-06 2012-07-04 上海交通大学 Automatic image semantic annotation method based on scale learning and correlated label dissemination

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
"Hidden-Concept Driven Multilabel Image Annotation and Label Ranking";Bing-Kun Bao et al.;《IEEE Transactions on Multimedia》;20120229;第14卷(第1期);第199-210页 *
"Latent Semantic Analysis-based Image Auto Annotation";Mahdia Bakalem et al.;《IEEE Conf. on Machine and Web Intelligence》;20101231;第2010年卷;第460-463页 *
"词间相关性的CMRM图像标注方法";刘咏梅等;《智能***学报》;20110831;第6卷(第4期);第350-354页 *

Also Published As

Publication number Publication date
CN103714178A (en) 2014-04-09

Similar Documents

Publication Publication Date Title
CN107861939B (en) Domain entity disambiguation method fusing word vector and topic model
Quan et al. Unsupervised product feature extraction for feature-oriented opinion determination
CN107133213B (en) Method and system for automatically extracting text abstract based on algorithm
CN105320960B (en) Voting-based cross-language subjective and objective emotion classification method
US10042896B2 (en) Providing search recommendation
WO2017166912A1 (en) Method and device for extracting core words from commodity short text
CN111241294A (en) Graph convolution network relation extraction method based on dependency analysis and key words
CN105808530B (en) Interpretation method and device in a kind of statistical machine translation
CN107590219A (en) Webpage personage subject correlation message extracting method
CN107391565B (en) Matching method of cross-language hierarchical classification system based on topic model
CN106778878B (en) Character relation classification method and device
CN103714178B (en) Automatic image marking method based on word correlation
US9262400B2 (en) Non-transitory computer readable medium and information processing apparatus and method for classifying multilingual documents
CN104199845B (en) Line Evaluation based on agent model discusses sensibility classification method
CN107818173B (en) Vector space model-based Chinese false comment filtering method
CN115630640A (en) Intelligent writing method, device, equipment and medium
CN111046171B (en) Emotion discrimination method based on fine-grained labeled data
CN106339371B (en) A kind of English-Chinese meaning of a word mapping method and device based on term vector
CN112101014A (en) Chinese chemical industry document word segmentation method based on mixed feature fusion
CN107451116B (en) Statistical analysis method for mobile application endogenous big data
Gong et al. A semantic similarity language model to improve automatic image annotation
CN110929518A (en) Text sequence labeling algorithm using overlapping splitting rule
CN108182443B (en) Automatic image labeling method and device based on decision tree
Xu et al. A classification of questions using SVM and semantic similarity analysis
CN109002540B (en) Method for automatically generating Chinese announcement document question answer pairs

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant