CN110046666B - Mass picture labeling method - Google Patents

Mass picture labeling method Download PDF

Info

Publication number
CN110046666B
CN110046666B CN201910312598.0A CN201910312598A CN110046666B CN 110046666 B CN110046666 B CN 110046666B CN 201910312598 A CN201910312598 A CN 201910312598A CN 110046666 B CN110046666 B CN 110046666B
Authority
CN
China
Prior art keywords
labeling
model
prediction
pictures
initial model
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910312598.0A
Other languages
Chinese (zh)
Other versions
CN110046666A (en
Inventor
何志权
许琦
何志海
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Deepvision Creative Technology Ltd
Original Assignee
Shenzhen Deepvision Creative Technology Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Deepvision Creative Technology Ltd filed Critical Shenzhen Deepvision Creative Technology Ltd
Priority to CN201910312598.0A priority Critical patent/CN110046666B/en
Publication of CN110046666A publication Critical patent/CN110046666A/en
Application granted granted Critical
Publication of CN110046666B publication Critical patent/CN110046666B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2411Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on the proximity to a decision surface, e.g. support vector machines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24147Distances to closest patterns, e.g. nearest neighbour classification

Landscapes

  • Engineering & Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a method for labeling a mass of pictures, which comprises the following steps: step 1, establishing an initial model according to labeling information of a plurality of pictures; step 2, predicting the unmarked picture by using the initial model; step 3, carrying out confidence coefficient analysis on the prediction result in the step 3, thereby selecting reliable prediction; step 4, replacing the prediction result of reliable prediction with labeling information; and 5, updating the initial model by using the marking information obtained in the step 4, and returning to the step 1 to repeat iteration. The method can quickly and effectively label massive pictures so as to solve the problem of the real bottleneck of data labeling in deep learning.

Description

Mass picture labeling method
Technical Field
The invention relates to the technical field of picture marking, in particular to a massive picture marking method.
Background
Deep learning is currently finding increasingly widespread use in both academic and industrial areas. However, the deep learning model has three resource limitation problems: the data size of model training is large, and if the model training depends on network transmission, the network bandwidth becomes a bottleneck; a large amount of computing resources, many deep learning model parameters and complex operation process result in huge computation amount; a large amount of annotation data is required to train complex models.
Annotating data is a tedious process and is typically performed point-by-point with a brush aimed at a target, as shown in fig. 1, where (a) is the original picture and (b) is the annotated picture. Therefore, the existing mode wastes time and labor, and in the labeling process, the labeling effect is often not ideal due to artificial factors and differences.
The mass pictures are marked by consuming a large amount of manpower and material resources, but a method for quickly and effectively marking the mass pictures does not exist at present.
Disclosure of Invention
The invention provides a method for labeling a mass of pictures, which aims to solve at least one technical problem.
In order to solve the above problem, as an aspect of the present invention, a method for labeling a large number of pictures is provided, including: step 1, establishing an initial model according to labeling information of a plurality of pictures; step 2, predicting the unmarked picture by using the initial model; step 3, carrying out confidence coefficient analysis on the prediction result in the step 3, thereby selecting reliable prediction; step 4, replacing the prediction result of reliable prediction with labeling information; and 5, updating the initial model by using the marking information obtained in the step 4, and returning to the step 1 to repeat iteration.
Preferably, the labeling information in step 1 is obtained by manual labeling.
Preferably, step 3 comprises: and setting a confidence coefficient function for evaluating the reliability of the prediction, wherein the function utilizes the original image information and the output vector of the softmax layer of the model to evaluate the prediction result.
Preferably, step 3 comprises: and judging the matching degree of each point in the prediction result by utilizing the pixels in the 4x4 field around the point to calculate the information such as the variance and the gradient of the contrast and the output vector of the softmax layer of the position model, and if the matching degree is matched according to the preset confidence threshold, the prediction is reliable.
Preferably, the anastomotic match is characterized by a neural network or a support vector machine.
Preferably, the updating the initial model by using the label information obtained in step 4 includes: let the initial model be θ t The last used model is θ' t-1 Then the new model is θ = α θ' t-1 +(1-α)θ t Where α is an exponential smoothing coefficient.
By adopting the technical scheme, the method can quickly and effectively label massive pictures to solve the real bottleneck problem of labeling data in deep learning, firstly labels a small number of pictures, trains a deep learning model, then predicts the remaining unmarked pictures by using the model, evaluates the prediction result, converts accurate prediction into labeling, and then optimizes the model, so that the accuracy of the model is higher and the labeling accuracy is higher and higher in circulating iteration.
Drawings
FIG. 1 is a schematic diagram illustrating a prior art labeling of a picture;
FIG. 2 schematically illustrates a flow chart of the present invention;
fig. 3 schematically shows a schematic diagram of selecting reliable prediction results according to confidence.
Detailed Description
Embodiments of the invention will be described in detail below with reference to the drawings, but the invention can be implemented in many different ways as defined and covered by the claims.
An initial model is assumed to exist, and if not, a small number of pictures can be artificially marked to train the initial model. For convenience of the following explanation, we take defect detection as an example, as shown in fig. 1.
Step 1: the unlabeled data is predicted with the initial model.
Step 2: and carrying out confidence degree analysis on the predicted result. Since the model performance of step 1 is not ideal at this time, the predicted result is good or bad, and therefore, the purpose of performing confidence analysis on the predicted result is to select reliable prediction.
Confidence function
Figure BDA0002031996550000031
To evaluate the reliability of model predictions. This function evaluates the curve part in (b) of fig. 1 using the original image information I and the output vector of the softmax layer of the model. Specifically, for each point on the curve, we calculate the variance, gradient, etc. of the contrast using the pixels in the 4 × 4 domain around the point, and at this position, the softmax layer output of the model represents the predicted defect strength. These two pieces of information, if matched with a high degree of agreement, indicate that the prediction is reliable. The mapping relationship can be characterized by a simple neural network or a support vector machine (support vector machine).
And step 3: and selecting a key sample. Step 2, confidence analysis is carried out on all predictions, and accordingly, reliable predictions can be selected. We can set a confidence threshold T, when the threshold is greater than T, indicating that the prediction is reliable, and leave, otherwise leave, as shown in fig. 3, the prediction confidence in the upper box in (b) is low, and leave in the lower box.
And 4, step 4: and converting the prediction result into the labeling information. The prediction result with high confidence and the artificial label are consistent on a large probability and can be converted into label information.
And 5: and updating the model. By using more labeled information, we can train out a model theta t The last used model is θ' t-1 Then the new model is θ = α θ' t-1 +(1-α)θ t Where α is an exponential smoothing coefficient. The model has better prediction performance than the previous model.
With the updated model, we return to step 1 to predict the unmarked part and update the existing prediction. The performance of the model is better and better by repeating the iteration, and the quality of the label is gradually improved.
The above description is only a preferred embodiment of the present invention and is not intended to limit the present invention, and various modifications and changes may be made by those skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (2)

1. A method for labeling massive pictures is characterized by comprising the following steps:
step 1, establishing an initial model according to labeling information of a plurality of pictures, wherein the labeling information is obtained in a manual labeling mode;
step 2, directly utilizing the initial model to predict the unmarked picture;
step 3, performing confidence analysis on the prediction result in the step 2, thereby selecting reliable predictions, which comprises the following steps: setting a confidence function for evaluating the reliability of prediction, wherein the function utilizes original image information and an output vector of a softmax layer of a model to evaluate a prediction result, judging the matching degree of each point in the prediction result between the variance and the gradient of the contrast ratio of each point and the output vector of the softmax layer of the model by utilizing pixels in the 4x4 field around each point, and if the point is matched according to a preset confidence threshold value, the prediction is reliable;
step 4, replacing the prediction result of reliable prediction with labeling information;
and 5, updating the initial model by using the marking information obtained in the step 4, wherein the method comprises the following steps: setting an initial model as
Figure DEST_PATH_IMAGE001
The last used model is
Figure DEST_PATH_IMAGE002
Then the new model is
Figure DEST_PATH_IMAGE003
Wherein
Figure DEST_PATH_IMAGE004
Is an exponential smoothing coefficient; then, the iteration is repeated by returning to the step 1.
2. The method for labeling the massive pictures according to claim 1, wherein the matching is characterized by being characterized by a neural network or a support vector machine.
CN201910312598.0A 2019-04-18 2019-04-18 Mass picture labeling method Active CN110046666B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910312598.0A CN110046666B (en) 2019-04-18 2019-04-18 Mass picture labeling method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910312598.0A CN110046666B (en) 2019-04-18 2019-04-18 Mass picture labeling method

Publications (2)

Publication Number Publication Date
CN110046666A CN110046666A (en) 2019-07-23
CN110046666B true CN110046666B (en) 2022-12-02

Family

ID=67277763

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910312598.0A Active CN110046666B (en) 2019-04-18 2019-04-18 Mass picture labeling method

Country Status (1)

Country Link
CN (1) CN110046666B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112699908B (en) * 2019-10-23 2022-08-05 武汉斗鱼鱼乐网络科技有限公司 Method for labeling picture, electronic terminal, computer readable storage medium and equipment
CN112101156A (en) * 2020-09-02 2020-12-18 杭州海康威视数字技术股份有限公司 Target identification method and device and electronic equipment

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4480156B2 (en) * 2005-02-02 2010-06-16 キヤノン株式会社 Image processing apparatus and method
CN106096627A (en) * 2016-05-31 2016-11-09 河海大学 The Polarimetric SAR Image semisupervised classification method that considering feature optimizes
CN107977667B (en) * 2016-10-21 2019-02-19 西安电子科技大学 SAR target discrimination method based on semi-supervised coorinated training
CN106897738B (en) * 2017-01-22 2019-07-16 华南理工大学 A kind of pedestrian detection method based on semi-supervised learning
US10210403B2 (en) * 2017-04-24 2019-02-19 Here Global B.V. Method and apparatus for pixel based lane prediction
CN108764281A (en) * 2018-04-18 2018-11-06 华南理工大学 A kind of image classification method learning across task depth network based on semi-supervised step certainly

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Restoration algorithm of damaged mark image based on texture synthesis;Min Qi et al.;《2015 International Conference on Computer and Computational Sciences (ICCCS)》;20151221;第1-4页 *

Also Published As

Publication number Publication date
CN110046666A (en) 2019-07-23

Similar Documents

Publication Publication Date Title
CN114241282B (en) Knowledge distillation-based edge equipment scene recognition method and device
CN109741332B (en) Man-machine cooperative image segmentation and annotation method
JP2018097807A (en) Learning device
US20050160340A1 (en) Resource-light method and apparatus for outlier detection
CN110046666B (en) Mass picture labeling method
US20220092407A1 (en) Transfer learning with machine learning systems
CN111581116B (en) Cross-project software defect prediction method based on hierarchical data screening
CN109919302B (en) Training method and device for neural network of image
CN112052818A (en) Unsupervised domain adaptive pedestrian detection method, unsupervised domain adaptive pedestrian detection system and storage medium
CN111160432A (en) Automatic classification method and system for panel production defects
CN114511077A (en) Training point cloud processing neural networks using pseudo-element based data augmentation
CN115277587B (en) Network traffic identification method, device, equipment and medium
CN115205727A (en) Experiment intelligent scoring method and system based on unsupervised learning
CN114782752B (en) Small sample image integrated classification method and device based on self-training
CN114330588A (en) Picture classification method, picture classification model training method and related device
CN111125389A (en) Data classification cleaning system and cleaning method based on dynamic progressive sampling
CN113223011B (en) Small sample image segmentation method based on guide network and full-connection conditional random field
CN117703023A (en) Seamless paving method for rock plates/tiles
CN117496426A (en) Precast beam procedure identification method and device based on mutual learning
CN110929792B (en) Image labeling method, device, electronic equipment and storage medium
CN108664902B (en) End-to-end video copy detection method and device based on deep learning
CN116089944A (en) Cross-platform application program abnormality detection method and system based on transfer learning
Sirhan et al. Multilabel CNN model for asphalt distress classification
CN114186090A (en) Intelligent quality inspection method and system for image annotation data
CN112164040A (en) Steel surface defect identification method based on semi-supervised deep learning algorithm

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant