CN103679218A - Handwritten form keyword detection method - Google Patents

Handwritten form keyword detection method Download PDF

Info

Publication number
CN103679218A
CN103679218A CN201310582398.XA CN201310582398A CN103679218A CN 103679218 A CN103679218 A CN 103679218A CN 201310582398 A CN201310582398 A CN 201310582398A CN 103679218 A CN103679218 A CN 103679218A
Authority
CN
China
Prior art keywords
keyword
image
moving window
unique point
point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201310582398.XA
Other languages
Chinese (zh)
Other versions
CN103679218B (en
Inventor
吕岳
张文超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
East China Normal University
Original Assignee
East China Normal University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by East China Normal University filed Critical East China Normal University
Priority to CN201310582398.XA priority Critical patent/CN103679218B/en
Publication of CN103679218A publication Critical patent/CN103679218A/en
Application granted granted Critical
Publication of CN103679218B publication Critical patent/CN103679218B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Character Discrimination (AREA)

Abstract

The invention discloses a handwritten form keyword detection method comprising that: characteristic points of keyword images in a keyword image database are extracted so that a keyword characteristic database is established; the characteristic points of texts in text images to be detected are extracted so that a characteristic point database of the text images to be detected is obtained; sliding windows of the text images to be detected are extracted, characteristic point sets corresponding to the sliding windows are extracted from the characteristic point database, and the characteristic point sets are compared so that initial matching point pair sets are obtained; the initial matching point pair sets are screened so that accurate matching point pairs are obtained; and the sliding windows in the text images to be detected are screened according to the matching point pairs and integrated so that detection results are obtained. In the process of characteristic extraction, SIFT characteristic representation is adopted. The method is applicable to detection of keywords of large amounts of handwritten form documents, such as historical literatures, correspondences, notes, etc. Under the premise of establishing a black list image database, the keywords in a black list can be effectively detected and handwriting of different authors can also be distinguished.

Description

A kind of handwritten form keyword spotting method
Technical field
The present invention relates to image detecting technique, relate in particular to a kind of specific people's of writing handwritten form keyword spotting method based on word local feature.
Background technology
Handwritten form keyword spotting is to detect specific keyword in a large amount of handwritten text images.The method of handwritten form keyword spotting is normally carried out character retrieval on the basis of word identification again at present.
Yet Chinese character classification is many, writing style is changeable, first the existing handwritten form keyword error-detecting method based on identification word needs to set up huge Chinese character template base, and spend the plenty of time and carry out training and the classification of feature, the image pre-service in early stage and Text segmentation also largely impact are known the other result of word, and then affect keyword spotting result; In addition the writing style that the method for the result of this dependence identification is not write difference people is included in and is considered, thus the identification to handwritten text effectively.
Summary of the invention
The present invention has overcome cannot identify the defect of the specific people's of writing handwritten text in prior art effectively, has proposed a kind of handwritten form keyword spotting method.
The present invention proposes a kind of handwritten form keyword spotting method, comprise the steps: step 1: obtain keyword image library, extract the unique point of keyword image in described keyword image library and set up keyword feature storehouse; Step 2: extract the unique point of text image Chinese version to be detected, obtain the unique point storehouse of described text image to be detected; Step 3: extract the moving window of text image to be detected, extract described moving window characteristic of correspondence point set from described unique point storehouse, by described unique point set and the contrast of described keyword feature storehouse, obtain initial matching double points set; Step 4: according to described moving window Chinese word geometry information, screen described initial matching double points set and obtain matching double points accurately; Step 5: moving window integration according in the described text image to be detected of described matching double points screening, obtain testing result.
In the handwritten form keyword spotting method that the present invention proposes, when the unique point in described step 3 in described unique point set and described keyword feature storehouse difference are less than threshold value, this unique point is detected as initial matching double points, and described unique point is shown below:
| &theta; s - &theta; w | < p 1 p 2 < &sigma; s &sigma; w < 1 p 2 | | d &RightArrow; s - d &RightArrow; w | | 2 < p 3 ;
In formula, θ sthe unique point deflection that represents keyword image, σ sthe unique point yardstick that represents keyword image,
Figure BDA0000417127290000012
the unique point eigenvector that represents keyword image, θ wthe unique point deflection that represents text image to be detected, σ wthe unique point yardstick that represents text image to be detected,
Figure BDA0000417127290000021
the unique point eigenvector that represents text image to be detected, p1, p2, p3 represent respectively threshold value.
In the handwritten form keyword spotting method that the present invention proposes, in described step 4, by the described initial matching double points set of word geometry information constraint screening, its process comprises the steps:
Step b1: set up the geological information constraints graph between described unique point set and the unique point set of described keyword image;
Step b2: search algorithm by Clique described geological information constraints graph is screened, delete the matching double points of mistake coupling in described initial matching double points set.
In the handwritten form keyword spotting method that the present invention proposes, the constraint condition of described geological information constraints graph is as shown in the formula expression:
|(x si-x sj)-(x wi-x sj)|<p4×Avg(σ s) ;
|(y si-y sj)-(y wi-y wj)|<p5×Avg(σ s)
In formula, x sthe horizontal ordinate of representation feature point in keyword image, x wthe horizontal ordinate of representation feature point in moving window image, y sthe ordinate of representation feature point in keyword image, y wthe ordinate of representation feature point in moving window image, Avg (σ s) representing the average characteristics point scale of keyword image, p4 and p5 represent threshold value.
In the handwritten form keyword spotting method that the present invention proposes, the moving window screening in described step 5 in described text image to be detected comprises the steps:
Step c1: if the centre-of gravity shift ratio of the center of gravity of the matching double points of described moving window and former keyword is greater than 0.15, delete this slip moving window, otherwise continue next step;
Step c2: if the match point of described moving window is greater than 0.18 in the left half of ratio distributing and the distribution proportion deviation of former keyword, deletes this slip moving window, otherwise retain this moving window.
This method will detect the specific people's of writing hand-written keyword by computed image matching degree, the keyword image of reserved particular person being write and text image to be detected carry out respectively feature extraction, use moving window to slide at text image to be detected, detect specific hand-written keyword.In characteristic extraction procedure, this method has adopted SIFT characteristic present.The method is applicable to the keyword spotting of a large amount of handwritten form documents, as historical document, letter, notes etc., is setting up under the prerequisite of blacklist image library, not only can effectively detect the keyword in blacklist, more can distinguish the person's handwriting of different authors.
Accompanying drawing explanation
Fig. 1 is the process flow diagram of handwritten form keyword spotting method of the present invention;
Fig. 2 is the schematic diagram of text image to be detected in the present embodiment;
Fig. 3 is the process flow diagram in keyword spotting stage in the present invention;
Fig. 4 is the schematic diagram of matching double points preliminary in the present embodiment;
Fig. 5 is the schematic diagram of geological information figure in the present embodiment;
Fig. 6 is the testing result schematic diagram of the present embodiment;
Fig. 7 is the schematic diagram of center of gravity in the testing result of the present embodiment;
Fig. 8 is the schematic diagram by mistake mating in the testing result of the present embodiment;
Fig. 9 is the testing result schematic diagram of the present embodiment.
Embodiment
In conjunction with following specific embodiments and the drawings, the present invention is described in further detail.Implement process of the present invention, condition, experimental technique etc., except the content of mentioning specially below, be universal knowledege and the common practise of this area, the present invention is not particularly limited content.
The flow process of this method as shown in Figure 1, is mainly divided into two stages: feature extraction phases and keyword spotting stage.In feature extraction phases, extract the unique point set of keyword image in keyword image library, and the unique point set of text image Chinese version to be detected, keyword feature point storehouse and unique point set to be detected after extraction, formed respectively.Keyword spotting in the stage compares both, thereby obtains the matching double points between text and keyword image, thereby obtains final matching double points by continuous screening, thereby completes the detection of keyword.
In feature extraction phases, first keyword image library is carried out to SIFT feature extraction in batches.Each envelope image can be described as the S set={ F of SIFT unique point i, each unique point
Figure BDA0000417127290000031
mainly 5 data, consist of, its concrete meaning is in Table 1.The unique point set of each keyword image is stored in text, forms feature database.
Table 1SIFT characteristic point data table
For text image to be detected, in advance it is gone and is cut apart.Document Segmentation to be detected is become to several moving windows, in each moving window, comprise several words.Preferably, when the text image of every a line being extracted to its SIFT unique point set, by the horizontal ordinate x value ascending order of unique point, arrange storage, be designated as
Figure BDA0000417127290000033
so that follow-up slip moving window extracts.
Due to the special repetitive structure of Chinese character, the direct matching result of keyword characteristics of image on text image is very poor, and what obtain is mistake matching characteristic point substantially.So text image is cut, and adopts slip moving window, characteristic matching is dwindled in the scope of several Chinese characters, can obtain more correct match point.
In the present embodiment, utilize set Wind (p, q) to represent moving window, wherein p and q represent respectively the origin coordinates of moving window and stop coordinate.If text image width is W d, moving window width is w, and moving window moving step length is s, and moving window quantity is n, as shown in Figure 2, according to following three kinds of methods, cuts apart text image and extracts slip moving window:
If 1 W d< w, n=1, p 1=0, q 1=W d;
2 otherwise, according to formula (1) below, calculate the coordinate of moving window
Figure BDA0000417127290000041
If 3 q i> W d, q i=W d, the not enough width of final stage also becomes a moving window.
After obtaining moving window set, each characteristic matching is only carried out for some moving windows, starts to right end, just as a moving window sliding from left to right from the left side of text image.Coupling in each moving window only drops on p for horizontal ordinate x iand q ibetween text feature point set.
In above formula, moving window width w and moving window moving step length s are respectively as following formula (2) represents:
w = W s &times; Avg ( &sigma; d ) Avg ( &sigma; s ) s = w 6 - - - ( 2 )
In formula, W sfor keyword picture traverse, Avg (σ s) be the average of keyword characteristics of image point scale, Avg (σ d) be the average of text image unique point.By unique point yardstick, regulate moving window width, can adapt to the dimensional variation of image.
The keyword spotting stage is the core of this method.The main process in this stage as shown in Figure 3.In each moving window, keyword image characteristic point can use Kd-tree algorithm find with moving window in the unique point of 5 arest neighbors in unique point set, form initial matching double points.
First utilize the information such as deflection, yardstick of unique point to carry out preliminary screening to matching double points, if do not consider rotation, distortion and the convergent-divergent of file and picture, so the deflection between two similar unique points, scale-value difference just can not be large especially.In addition, the king-sized matching double points of the Euclidean distance of eigenvector also can be deleted, because eigenvector representation feature point gradient statistical information around.Utilize formula (3) to matching characteristic point to carrying out preliminary screening
| &theta; s - &theta; w | < p 1 p 2 < &sigma; s &sigma; w < 1 p 2 | | d &RightArrow; s - d &RightArrow; w | | 2 < p 3 - - - ( 3 )
In formula, θ s, σ s,
Figure BDA0000417127290000052
and θ w, σ w, the unique point deflection, yardstick, the eigenvector that represent respectively keyword image and moving window image.P1, p2, p3 are respectively corresponding threshold values.The result of match point after preliminary screening as shown in Figure 4
Fig. 4 has intercepted the fragment of slip moving window in text image, the position that rectangle frame is moving window, and what show below is the image of keyword, thin straight line has connected the unique point of mating in two width images.
In repetitive structure Fig. 5 due to Chinese character, still there are a lot of Mismatching points.Utilize unique point to set up geometrical constraint in the space distribution of word, can further delete Mismatching point pair.The first step is first set up geometrical constraint figure, and its step is as follows:
1, establish S={s 1..., s mbe keyword characteristics of image point set, W={w 1..., w mit is the feature point set matching in moving window image.Non-directed graph G=(V, E) represents geometrical constraint figure.V={v 1... v mbe the vertex set of G, wherein v i=(s i, w i), represent in S and W that every a pair of match point is with regard to a summit in corresponding G.
Figure BDA0000417127290000054
set for G limit.
2, add as follows the limit of G, for any two unique points to v i=(s i, w i) and v j=(s j, w j), meet constraint condition
|(x si-x sj)-(x wi-x wj)|<p4×Avg(σ s) (4)
|(y si-y sj)-(y wi-y wj)|<p5×Avg(σ s)
In vertex v iand v jbetween add a limit.X in above formula s, x wthe horizontal ordinate of representation feature point in keyword image and moving window image, and y sand y wrepresent its ordinate, with reference to figure 5 examples.P4 and p5 are two threshold values, in order to adapt to the size of different keyword images, be multiplied by the average dimension value of keyword image characteristic point.Although geometrical constraint figure is the method for setting up image geometry contact of relatively commonly using, different constraint condition can build different constraints graphs, but unique constraint condition that the present invention proposes, as shown in formula (4), the architectural feature that meets word, be suitable for the screening that unique point is right, to improving accuracy rate of the present invention, have very large effect.
After more than setting up geometrical constraint figure, suppose wherein all unique points all correctly coupling, in GCG figure, all summits will be connected to each other so, form agglomerate closely.When existence is mated by mistake, will in GCG figure, find a clique, in this subgraph, all summits are all interconnected, can search algorithm with Clique and realize, its concrete steps are as follows:
1, initialization candidate vertices collection C=V, Clique vertex set M=φ;
2, calculate the degree on each summit in V, be designated as set Deg (V);
3, select the maximum vertex v of Deg (V) in C 0, added M, i.e. M=M ∪ { v 0;
4, by v 0from C, delete, and only in C, retain and v 0the summit being connected, C=C v 0, C=C ∩ N (v 0).N (v wherein 0) expression v 0adjacent vertex collection.
If 5 C=φ, algorithm finishes, otherwise forwards 3 repetition said process to.
By setting up geometrical constraint figure and Clique, search algorithm, can delete a large amount of Mismatching points pair, result is as shown in Figure 6 after geometrical constraint is processed for the initial matching in Fig. 4.Contrast the visible matching double points that many mistakes are mated of having deleted with Fig. 4.See the contrast of Fig. 8 and Fig. 9, moving window has covered " impelling " two words again, and due to the similarity of some strokes, this moving window and keyword image " food " have produced a large amount of error matching points pair.According to text geometry information delete Mismatching point to after, the matching double points showing in Fig. 9 has just reduced a lot, can find out that it is different character images that the present invention can judge.Geometrical constraint figure has effectively utilized the structural information of word, can greatly improve the correctness of feature point pair matching, plays in the methods of the invention central role.
By geometrical constraint to after the screening of match point, the quantity that each moving window can a corresponding match point, the very few moving window of quantity is deleted.Because the complicated and simple degree difference of Chinese character causes the unique point quantity of himself also different, make the match point quantity variance of different Chinese character very large, therefore adopt adaptive threshold to screen moving window.Formula (5) shows to adjust threshold value according to the quantity of SIFT unique point in keyword image
tn=k×S fnum (5)
In formula, tn is threshold value, S fnumfor keyword characteristics of image, count, k gets empirical value 0.25.
When the some words in keyword appear in text image, often the point of coupling is also more.This method adopts two kinds of strategies to judge: the one, and the judgement based on centre-of gravity shift, the 2nd, the judgement based on a distribution.
When the set of a certain moving window unique point is mated with keyword image characteristic point, can obtain mating two set of point set after initial matching point set and screening, if while being greater than threshold value through remaining the center of gravity horizontal ordinate of match point and the coordinate offset amount of corresponding keyword image characteristic point after Mismatching point screening, delete this moving window; If while being greater than threshold value through remaining the distribution proportion that (comprises left side or right side) in keyword image of match point and the distribution proportion of keyword image characteristic point in keyword image after Mismatching point screening, delete this moving window.When meeting following formula
|XCent(S 0)-XCent(S x)|>tc (6)
|LPoi(S 0)-LPoi(S x)|>tp
Time, current moving window is deleted.XCent in formula (S) represents the center of gravity horizontal ordinate of keyword image characteristic point, and LPoi (S) represents that keyword image characteristic point is at the distributive law of left one side of something.Tc and tp are two threshold values, get empirical value 0.15 and 0.18.As shown in Figure 7, two black circles represent the center of gravity of keyword image, from middle position, moved to the position in left side, because after screening Mismatching point, match point has substantially all concentrated on Chinese character " food ", although match point number is also many, but in the time of can finding out the judgement through centre-of gravity shift, distribution and the keyword image deviations of its match point are very large, this moving window still can be deleted, " property food " can not be detected as " food " yet, thereby reduced the error rate of handwritten form identification.
Finally need remaining moving window to merge, because some overlapping moving window has all comprised same keyword.When two moving window laps account for the more than 60% of whole moving window, they are merged into a moving window.Moving window after merging is exactly the candidate keywords moving window finally detecting.
The present embodiment is tested in 2764 envelope text images, by 50 authors, is write, and everyone 50 texts are not etc.Extract 15 of keywords, 99 keyword to be detected positions, carry out 10 groups of experiments.Use recall rate R, the alert rate E of mistake and new F value evaluation experimental result.Suppose detected image W envelope, S keyword detected altogether, wherein correctly detect Y, flase drop N, S=Y+N, the keyword that should detect has T, and above-mentioned three judgment criteria are calculated according to formula (7)
R = P T E = N W F = 2 0.8 &times; 1 R + 1.2 &times; ( 1 1 - E ) - - - ( 7 )
By formula (7), visible R can embody this method to keyword spotting ability, more high better; E will try one's best lowly, the least possiblely in to the detection of great amount of images must introduce erroneous judgement, embodies the distinguishing ability of the method to keyword.New F value is the balance to above two standards, be R and E according to the harmonic-mean of different weight calculation, concrete outcome is in Table 2.
Table 2 the present embodiment keyword spotting result
R E F
Optimum value 91.92% 0.65% 94.91%
Worst-case value 85.86% 3.12% 93.18%
Mean value 88.64% 1.85% 94.09%
Protection content of the present invention is not limited to above embodiment.Do not deviating under the spirit and scope of inventive concept, variation and advantage that those skilled in the art can expect are all included in the present invention, and take appending claims as protection domain.

Claims (5)

1. a handwritten form keyword spotting method, is characterized in that, comprises the steps:
Step 1: obtain keyword image library, extract the unique point of keyword image in described keyword image library and set up keyword feature storehouse;
Step 2: extract the unique point of text image Chinese version to be detected, obtain the unique point storehouse of described text image to be detected;
Step 3: extract the moving window of text image to be detected, extract described moving window characteristic of correspondence point set from described unique point storehouse, by described unique point set and the contrast of described keyword feature storehouse, obtain initial matching double points set;
Step 4: according to described moving window Chinese word geometry information, screen described initial matching double points set and obtain matching double points accurately;
Step 5: moving window integration according in the described text image to be detected of described matching double points screening, obtain testing result.
2. handwritten form keyword spotting method as claimed in claim 1, it is characterized in that, when the unique point in described step 3 in described unique point set and described keyword feature storehouse difference are less than threshold value, this unique point is detected as initial matching double points, and described unique point is shown below:
| &theta; s - &theta; w | < p 1 p 2 < &sigma; s &sigma; w < 1 p 2 | | d &RightArrow; s - d &RightArrow; w | | 2 < p 3 ;
In formula, θ sthe unique point deflection that represents keyword image, σ sthe unique point yardstick that represents keyword image,
Figure FDA0000417127280000012
the unique point eigenvector that represents keyword image, θ wthe unique point deflection that represents text image to be detected, σ wthe unique point yardstick that represents text image to be detected,
Figure FDA0000417127280000013
the unique point eigenvector that represents text image to be detected, p1, p2, p3 represent respectively threshold value.
3. handwritten form keyword spotting method as claimed in claim 1, is characterized in that, in described step 4, by the described initial matching double points set of word geometry information constraint screening, its process comprises the steps:
Step b1: set up the geological information constraints graph between described unique point set and the unique point set of described keyword image;
Step b2: search algorithm by Clique described geological information constraints graph is screened, delete the matching double points of mistake coupling in described initial matching double points set.
4. handwritten form keyword spotting method as claimed in claim 3, is characterized in that, the constraint condition of described geological information constraints graph is as shown in the formula expression:
|(x si-x sj)-(x wi-x wj)|<p4×Avg(σ s)
|(y si-y sj)-(y wi-y wj)|<p5×Avg(σ s)
In formula, x sthe horizontal ordinate of representation feature point in keyword image, x wthe horizontal ordinate of representation feature point in moving window image, y sthe ordinate of representation feature point in keyword image, y wthe ordinate of representation feature point in moving window image, Avg (σ s) representing the average characteristics point scale of keyword image, p4 and p5 represent threshold value.
5. handwritten form keyword spotting method as claimed in claim 1, is characterized in that, the moving window screening in described step 5 in described text image to be detected comprises the steps:
Step c1: if the centre-of gravity shift ratio of the center of gravity of the match point of described moving window and former keyword is greater than 0.15, deleting should
Slip moving window, otherwise continue next step;
Step c2: if the match point of described moving window is greater than 0.18 in the left half of ratio distributing and the distribution proportion deviation of former keyword, deletes this slip moving window, otherwise retain this moving window.
CN201310582398.XA 2013-11-19 2013-11-19 A kind of handwritten form keyword detection method Active CN103679218B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310582398.XA CN103679218B (en) 2013-11-19 2013-11-19 A kind of handwritten form keyword detection method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310582398.XA CN103679218B (en) 2013-11-19 2013-11-19 A kind of handwritten form keyword detection method

Publications (2)

Publication Number Publication Date
CN103679218A true CN103679218A (en) 2014-03-26
CN103679218B CN103679218B (en) 2017-01-04

Family

ID=50316706

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310582398.XA Active CN103679218B (en) 2013-11-19 2013-11-19 A kind of handwritten form keyword detection method

Country Status (1)

Country Link
CN (1) CN103679218B (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104408191A (en) * 2014-12-15 2015-03-11 北京国双科技有限公司 Method and device for obtaining correlated keywords of keywords
CN110851605A (en) * 2019-11-14 2020-02-28 携程计算机技术(上海)有限公司 Detection method and system for image-text information matching of OTA hotel and electronic equipment
CN111931508A (en) * 2020-08-24 2020-11-13 上海携旅信息技术有限公司 Digital conversion method and system, text processing method and system, device and medium
CN112150464A (en) * 2020-10-23 2020-12-29 腾讯科技(深圳)有限公司 Image detection method and device, electronic equipment and storage medium
CN112199545A (en) * 2020-11-23 2021-01-08 湖南蚁坊软件股份有限公司 Keyword display method and device based on picture character positioning and storage medium

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2166488A2 (en) * 2008-09-18 2010-03-24 Xerox Corporation Handwritten word spotter using synthesized typed queries
CN101819680A (en) * 2010-05-12 2010-09-01 上海交通大学 Detection method of picture matching point pair

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP2166488A2 (en) * 2008-09-18 2010-03-24 Xerox Corporation Handwritten word spotter using synthesized typed queries
CN101819680A (en) * 2010-05-12 2010-09-01 上海交通大学 Detection method of picture matching point pair

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
PENG WANG ET AL.: "Exploring Interest Points and Local Descriptors", 《15TH INTERNATIONAL CONFERENCE, CAIP 2013》, 31 August 2013 (2013-08-31), pages 408 - 415, XP047038805, DOI: doi:10.1007/978-3-642-40246-3_51 *
张洁玉等: "基于SIFT特征点匹配的印刷品图像检测方法", 《江南大学学报(自然科学版)》, vol. 6, no. 6, 31 December 2007 (2007-12-31), pages 850 - 854 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104408191A (en) * 2014-12-15 2015-03-11 北京国双科技有限公司 Method and device for obtaining correlated keywords of keywords
CN104408191B (en) * 2014-12-15 2017-11-21 北京国双科技有限公司 The acquisition methods and device of the association keyword of keyword
CN110851605A (en) * 2019-11-14 2020-02-28 携程计算机技术(上海)有限公司 Detection method and system for image-text information matching of OTA hotel and electronic equipment
CN111931508A (en) * 2020-08-24 2020-11-13 上海携旅信息技术有限公司 Digital conversion method and system, text processing method and system, device and medium
CN111931508B (en) * 2020-08-24 2023-05-12 上海携旅信息技术有限公司 Digital conversion method and system, text processing method and system, equipment and medium
CN112150464A (en) * 2020-10-23 2020-12-29 腾讯科技(深圳)有限公司 Image detection method and device, electronic equipment and storage medium
CN112150464B (en) * 2020-10-23 2024-01-30 腾讯科技(深圳)有限公司 Image detection method and device, electronic equipment and storage medium
CN112199545A (en) * 2020-11-23 2021-01-08 湖南蚁坊软件股份有限公司 Keyword display method and device based on picture character positioning and storage medium
CN112199545B (en) * 2020-11-23 2021-09-07 湖南蚁坊软件股份有限公司 Keyword display method and device based on picture character positioning and storage medium

Also Published As

Publication number Publication date
CN103679218B (en) 2017-01-04

Similar Documents

Publication Publication Date Title
CN104766084B (en) A kind of nearly copy image detection method of multiple target matching
CN105205488B (en) Word area detection method based on Harris angle points and stroke width
CN103679218A (en) Handwritten form keyword detection method
Pal et al. Touching numeral segmentation using water reservoir concept
CN103049750B (en) Character identifying method
CN103914680B (en) A kind of spray printing character picture identification and check system and method
CN101719142B (en) Method for detecting picture characters by sparse representation based on classifying dictionary
CN101299236B (en) Method for recognizing Chinese hand-written phrase
CN103258037A (en) Trademark identification searching method for multiple combined contents
CN103593464A (en) Video fingerprint detecting and video sequence matching method and system based on visual features
CN103473545B (en) A kind of text image method for measuring similarity based on multiple features
CN104715254A (en) Ordinary object recognizing method based on 2D and 3D SIFT feature fusion
CN102388392A (en) Pattern recognition device
CN108520254A (en) A kind of Method for text detection, device and relevant device based on formatted image
Miller et al. A set of handwriting features for use in automated writer identification
Li et al. Detecting text lines in handwritten documents
CN104408711A (en) Multi-scale region fusion-based salient region detection method
CN106127243A (en) A kind of image matching method describing son based on binaryzation SIFT
CN104951788A (en) Extracting method of strokes of separate character in calligraphy work
Sharma et al. High‐level feature aggregation for fine‐grained architectural floor plan retrieval
CN107679401A (en) A kind of malicious web pages recognition methods and device
CN104835174A (en) Robustness model fitting method based on supermap mode search
CN106503706B (en) The method of discrimination of Chinese character pattern cutting result correctness
Thean et al. Textual summarisation of flowcharts in patent drawings for CLEF-IP 2012
Bhattacharya et al. Overwriting repetition and crossing-out detection in online handwritten text

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant