CN113096079A - Image analysis system and construction method thereof - Google Patents

Image analysis system and construction method thereof Download PDF

Info

Publication number
CN113096079A
CN113096079A CN202110338180.4A CN202110338180A CN113096079A CN 113096079 A CN113096079 A CN 113096079A CN 202110338180 A CN202110338180 A CN 202110338180A CN 113096079 A CN113096079 A CN 113096079A
Authority
CN
China
Prior art keywords
image
block
network
training
unit
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202110338180.4A
Other languages
Chinese (zh)
Other versions
CN113096079B (en
Inventor
廖欣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
West China Second University Hospital of Sichuan University
Original Assignee
West China Second University Hospital of Sichuan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by West China Second University Hospital of Sichuan University filed Critical West China Second University Hospital of Sichuan University
Priority to CN202110338180.4A priority Critical patent/CN113096079B/en
Publication of CN113096079A publication Critical patent/CN113096079A/en
Application granted granted Critical
Publication of CN113096079B publication Critical patent/CN113096079B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/55Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02TCLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
    • Y02T10/00Road transport of goods or passengers
    • Y02T10/10Internal combustion engine [ICE] based vehicles
    • Y02T10/40Engine management systems

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Databases & Information Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides an image analysis system and a construction method thereof. The system comprises an image database construction unit, a convolutional neural network unit and an analysis unit; the image database construction unit comprises an image data acquisition unit, an image data annotation unit and an image database construction unit; the convolutional neural network unit comprises a convolutional neural network model construction unit and a convolutional neural network model training unit; and the analysis unit analyzes the specific image structure in the image to be analyzed by using the trained image anomaly detection model. The image analysis system is simple in structure, can quickly identify the image and output an analysis result, and improves the judgment accuracy, the working efficiency and the working continuous state.

Description

Image analysis system and construction method thereof
Technical Field
The invention relates to the field of image analysis, in particular to an image analysis system and a construction method thereof.
Background
At present, with the research and progress of artificial intelligence technology, the artificial intelligence technology is being applied to various fields, the artificial intelligence technology is a comprehensive subject and relates to a wide range of fields, and the technology of the existing hardware level and the technology of the software level are also available. The artificial intelligence infrastructure generally includes technologies such as sensors, dedicated artificial intelligence chips, cloud computing, distributed storage, big data processing technologies, operation/interaction systems, mechatronics, and the like. The artificial intelligence software technology mainly includes computer vision technology, voice processing technology, natural language processing technology, Machine Learning (Machine Learning)/Deep Learning (Deep Learning), and the like.
Machine learning specializes in studying how computers simulate or implement human learning behaviors to acquire new knowledge or skills and reorganize existing knowledge structures to improve their performance. Machine learning is the core of artificial intelligence, is the fundamental approach for computers to have intelligence, and is applied to all fields of artificial intelligence. At present, machine learning models in various forms have thoroughly changed many fields of artificial intelligence, and are widely applied to intelligent analysis and recognition of images.
Disclosure of Invention
In order to overcome the defects in the prior art, the invention aims to provide an image analysis system and a construction method thereof.
In order to achieve the above object of the present invention, the present invention provides an image analysis system, comprising an image database construction unit, a convolutional neural network unit, and an analysis unit;
the image database construction unit comprises an image data acquisition unit for acquiring input image data, an image data labeling unit for labeling different image structures in each input image data, and an image database construction unit for classifying and sorting the labeled image data provided by the image data labeling unit;
the convolutional neural network unit comprises a convolutional neural network model construction unit and a convolutional neural network model training unit, and the convolutional neural network model construction unit is used for constructing an image anomaly detection model; the convolutional neural network model training unit trains the image anomaly detection model, inputs the model into a training image set, loss function weights, a feature extraction network and a classification network, and outputs a feature set S comprising the training set, a trained feature extraction network and weights f thereofθ
And the analysis unit analyzes the specific image structure in the image to be analyzed by using the trained image anomaly detection model.
The image analysis system disclosed by the invention is simple in structure, can be used for rapidly identifying the image and outputting an analysis result, and can be used for improving the judgment accuracy, the working efficiency and the working continuous state.
In the preferable scheme of the image analysis system, the convolutional neural network construction unit comprises a feature extraction network and a classification network;
the feature extraction network comprises a lower layer formed by M convolution modules BLOCK-A, a higher layer formed by N residual convolution modules BLOCK-B, P residual convolution modules BLOCK-C, a convolution layer connected with the higher layer and a tanh () activation function;
the convolution module BLOCK-A consists of a convolution layer and a LeakyReLU () activation function;
the residual convolution module BLOCK-B is composed of superposed convolution kernels of 1x1 and 3x3 and a skip layer;
the residual convolution module BLOCK-C is composed of a convolution kernel and a layer jump, wherein the convolution kernel is superposed with 1x1, 1x3 and 3x 1;
the classification network consists of a full connectivity layer and a LeakyReLU () activation function.
The feature extraction network has a plurality of convolution kernels with different scales on the same layer, so that sparse and non-sparse features can be learned at the same time, and layer jump ensures that deep and shallow network features can be considered by the network at the same time. The two characteristics in the network structure design increase the feature expression capability of the network.
In the preferable scheme of the image analysis system, the convolutional neural network construction unit comprises a feature extraction network and a classification network;
the feature extraction network comprises a lower layer formed by K convolution modules BLOCK-D, a higher layer formed by Q self-supervision convolution modules BLOCK-F, a convolution layer connected with the higher layer and a tanh () activation function;
the convolution module BLOCK-D consists of a convolution layer and a LeakyReLU () activation function;
the self-supervision convolution module BLOCK-F module comprises a plurality of 1x1 and 3x3 convolution kernels and a mean pooling layer which are mutually overlapped;
the classification network consists of a full connectivity layer and a LeakyReLU () activation function.
The feature extraction network enables the same layer to have convolution kernels with various scales, and can learn sparse and non-sparse features simultaneously, so that the feature expression capability of the network is improved.
In the preferable scheme of the image analysis system, the convolutional neural network construction unit comprises a feature extraction network and a classification network;
the feature extraction network comprises a plurality of convolution layers, and a BLOCK-G module is introduced in the middle of the convolution layers;
the BLOCK-G module comprises a plurality of superimposed 1x1, 3x3, 5x5 convolution kernels and a max pooling layer;
the classification network consists of a full connectivity layer and a LeakyReLU () activation function.
The feature extraction network of the invention enables the same layer to have convolution kernels with various scales, thereby increasing the feature expression capability of the network.
The application also provides a construction method of the image analysis system, which comprises the steps of constructing the image analysis system, obtaining an original image by an image database construction unit, labeling different image structures in the original image, classifying and sorting the labeled image data, and dividing a training set, a check set and a test set; training the image anomaly detection model by adopting a training set to obtain an ideal image anomaly detection model, and verifying the ideal image anomaly detection model by adopting a verification set to detect the accuracy of the ideal image anomaly detection model; testing an ideal image anomaly detection model by adopting a test set to detect the robustness of the ideal image anomaly detection model;
if the difference between the accuracy of the image anomaly detection model on the test set and the accuracy in the training of the check set exceeds a preset value, the model is over-fitted, the model returns to the convolutional neural network training unit, the network structure or the parameters are adjusted to perform retraining again, so that the image anomaly detection model with the difference between the accuracy on the test set and the accuracy in the training of the check set within the preset value is obtained, and at the moment, the robustness of the image anomaly detection model is high.
According to the system construction method, due to the introduction of the check set and the test set, the phenomena of under-fitting and over-fitting of the ideal weight of the abnormal detection model can be avoided, and the robustness of the ideal weight of the abnormal detection model obtained through training is guaranteed.
The preferable scheme of the image analysis system construction method is that the training flow of the training set to the image anomaly detection model is as follows:
b1: training a feature extraction network of an image anomaly detection model by using images in a training set;
b2: using the trained feature extraction network to obtain and store a feature set S, S ← S { (f) } corresponding to the training setθ(p) }, i.e.: randomly extracting an image block aiming at each image in the training set, wherein the image block has the same receiving field as the feature extraction network, and acquiring a feature vector f of the image block through the feature extraction network obtained by trainingθ(p) the feature vectors as a whole constitute a feature vector set S;
b3: saving trained feature extraction network weights fθAnd a feature vector set S corresponding to the training set.
The method and the device obtain the ideal weight after the training of the anomaly detection model is finished, and obtain the image characteristic set of the specified image structure according to the ideal weight and the images of the training set, so that the method and the device are used for realizing the anomaly detection of subsequent test images, can quickly identify the images and output the analysis result, and improve the judgment accuracy.
In a preferred embodiment of the image analysis system construction method, the step B1 includes the following steps:
b11: for each image in the training set, randomly selecting an image block p in an eight-neighborhood of a 3x3 grid of the image, wherein the scale of the image block p is the same as the receiving field of the feature extraction network, then randomly dithering the center of the image block p to obtain an image block p1, calculating the cross entropy of the image block p and the image block p1 as a sub-term Loss function Loss _1,
Figure BDA0002998437320000051
wherein the image block p1Relative to pThe true relative positions are y {0,1, …,7}, yiReferring to the number of image blocks of the 8 relative positions of the category i in y {0,1, …,7} in the training set; classifier CφIs trained to correctly predict image block p1Relative to the image block p, i.e. y ═ Cφ(fθ(p),fθ(p1)),aiThe confidence of the category i calculated by the classifier, and N is the total number of samples in the training set;
for an image block p, randomly selecting an image block p which is in the same row or column but not adjacent to the image block p in the four neighborhoods of the 5 × 5 grid2,p2The scale is the same as the acceptance field of the feature extraction network, the cross entropy of the image blocks p and p2 is calculated as a sub-term Loss function Loss _2,
Figure BDA0002998437320000052
wherein the image block p2The true relative position with respect to p is y {0,1,2,3}, yiReferring to the number of image blocks of the 4 relative positions of the category i in the training set, namely y {0,1,2,3 }; classifier CφIs trained to correctly predict image block p2Relative to the image block p, i.e. y ═ Cφ(fθ(p),fθ(p2)),biIs the confidence of class i calculated by the classifier;
aiming at an image block p, acquiring 2-4 image blocks p3, p4, p5 and p6 of four adjacent intersection areas of p, calculating L2 norm distances between p and selected image blocks p3, p4, p5 and p6, and averaging the distances to be used as a subitem Loss function Loss _3,
Figure BDA0002998437320000061
||fθ(p)-fθ(p2+i)||2the L2 norm distance between the image block p and a selected image block in p3, p4, p5 and p 6;
b12: calculating Loss function Loss of network model as lambda1*Loss_1+λ2*Loss_2+Loss_3,λ1、λ2The weight value in the loss function is larger than 0, and the Adam optimizer is utilized to carry out back propagation so as to realize the network weight of the feature extraction network modelRepeating iteration and optimizing;
b13: and C, repeatedly executing the steps B11-B12 until the number of turns is specified, and selecting and storing the optimal weight of the feature extraction network and the classification network according to the loss function of each turn of training. In a preferred embodiment of the image analysis system construction method, the step B1 includes the following steps:
step 1: for each image in the training set, randomly selecting an image block p7 in the eight neighborhood of the 3x3 grid of the image block p, calculating the cross entropy of the image blocks p and p7 as a sub-item Loss function Loss _4,
Figure BDA0002998437320000062
wherein the image block p7The true relative position with respect to p is y {0,1, …,7}, yiReferring to the number of image blocks of the 8 relative positions of the category i in y {0,1, …,7} in the training set; classifier CφIs trained to correctly predict image block p7Relative to the image block p, i.e.
Figure BDA0002998437320000064
ciIs the probability value of the category i calculated by the classifier, and N is the total number of samples in the training set;
for an image block p, randomly taking an image block p8 which is in the same row or column but not adjacent to the image block p in the four neighborhoods of the 5X5 network, calculating the cross entropy of p and p8 as a sub-term Loss function Loss _5,
Figure BDA0002998437320000063
wherein the image block p8The true relative position with respect to p is y {0,1,2,3}, yiReferring to the number of image blocks of the 4 relative positions of the category i in the training set, namely y {0,1,2,3 }; classifier CφIs trained to correctly predict image block p8Relative to the image block p, i.e.
Figure BDA0002998437320000065
diIs the probability value of the category i calculated by the classifier;
step 2: calculating a Loss function Loss of the network model, wherein lambda is a weight value in the Loss function and is larger than 0, and performing back propagation by using an Adam optimizer to realize network weight iteration and optimization of the feature extraction network model;
and step 3: and (3) repeatedly executing the steps 1-2 to designate the number of rounds, and selecting and storing the optimal weights of the feature extraction network and the classification network according to the loss function of each round of training.
The training process of the invention solves the problems that the number of the feature centers of the complex image is uncertain, and the workload of distributing corresponding image blocks to different feature centers is extremely large. Because the image blocks randomly selected directly from the training images have larger intra-class variance variation, part of the image blocks correspond to the background and part of the image blocks contain the target, and the situation that the image blocks contain both the background and the target can also exist. Therefore, mapping all features of different image blocks to one center, performing unimodal clustering, will weaken the link between features and content. In order to solve the problem, the scheme does not clearly define a center and divide corresponding image blocks, on the contrary, spatial adjacent image blocks are sampled to obtain image blocks with similar semantics, then a feature extraction network is trained to automatically collect the image blocks with similar features and semantics, and when the trained feature extraction network can well solve the pre-task, the network is considered to be capable of extracting effective features.
The system construction method can select the optimal training result from the results of the appointed training rounds of the anomaly detection model according to the loss function, and obtain the ideal weight of the anomaly detection model. Due to the design of the deep anomaly detection network structure and the self-supervision learning technology introduced in the construction of the loss function, the method can complete model training under the condition of a small sample data set, and further realize the analysis work of the target image structure.
The invention has the beneficial effects that: the invention has the advantages of high accuracy, short time consumption, long work duration and wide application range, can be widely applied to the fields of medical treatment, traffic safety and the like, particularly in the medical treatment field, is beneficial to solving the problem of uneven medical resource distribution, can realize remote high-quality medical treatment and the like, and provides more convenient and more accurate pathological diagnosis service for patients.
Additional aspects and advantages of the invention will be set forth in part in the description which follows and, in part, will be obvious from the description, or may be learned by practice of the invention.
Drawings
The above and/or additional aspects and advantages of the present invention will become apparent and readily appreciated from the following description of the embodiments, taken in conjunction with the accompanying drawings of which:
FIG. 1 is a schematic diagram of the structure of an image analysis system;
FIG. 2 is a schematic diagram of a first convolutional neural network construction unit network structure;
FIG. 3 is a schematic diagram of a convolution module BLOCK-A network structure;
FIG. 4 is a schematic diagram of a convolution module BLOCK-B network structure;
FIG. 5 is a schematic diagram of a convolution module BLOCK-C network structure;
FIG. 6 is a schematic diagram of a second convolutional neural network construction unit network structure;
FIG. 7 is a schematic diagram of a convolution module BLOCK-D network structure;
FIG. 8 is a schematic diagram of a convolution module BLOCK-F network structure;
FIG. 9 is a schematic diagram of a third convolutional neural network construction unit network structure;
FIG. 10 is a schematic diagram of a convolution module BLOCK-G network structure;
FIG. 11 is a schematic diagram of eight neighborhoods of a 3 × 3 grid of image blocks p;
FIG. 12 is a diagram of four neighborhood domains of a 5 × 5 grid of image blocks p;
fig. 13 is a schematic diagram of four adjacent intersection areas of an image block p.
Detailed Description
Reference will now be made in detail to embodiments of the present invention, examples of which are illustrated in the accompanying drawings, wherein like or similar reference numerals refer to the same or similar elements or elements having the same or similar function throughout. The embodiments described below with reference to the accompanying drawings are illustrative only for the purpose of explaining the present invention, and are not to be construed as limiting the present invention.
In the description of the present invention, unless otherwise specified and limited, it is to be noted that the terms "mounted," "connected," and "connected" are to be interpreted broadly, and may be, for example, a mechanical connection or an electrical connection, a communication between two elements, a direct connection, or an indirect connection via an intermediate medium, and specific meanings of the terms may be understood by those skilled in the art according to specific situations.
As shown in fig. 1, the present invention provides an image analysis system, which includes an image database construction unit, a convolutional neural network unit, and an analysis unit.
The image database construction unit comprises an image data acquisition unit, an image data annotation unit and an image database construction unit. The image data acquisition unit is used for acquiring input image data, the image data labeling unit is used for labeling different image structures in each input image data, and the image database construction unit is used for classifying and sorting the labeled image data provided by the image data labeling unit, dividing a training set, a check set and a test set and constructing an image database.
The convolutional neural network unit comprises a convolutional neural network model construction unit and a convolutional neural network model training unit. The convolutional neural network model construction unit is used for constructing an image anomaly detection model; the convolutional neural network model training unit trains the image anomaly detection model to obtain an ideal image anomaly detection model, the input of the convolutional neural network model training unit comprises a training image set, a loss function weight parameter, a feature extraction network and a classification network, and the output of the convolutional neural network model training unit comprises a feature set S of the training set, the trained feature extraction network and a weight f thereofθ
And the analysis unit analyzes the specific image structure in the image to be analyzed by using the trained image anomaly detection model.
In addition, the input terminal is used for inputting the existing image into the image data obtaining unit, and the input data is finally classified and collected by the image database construction unit for supporting the subsequent image analysis work. The output terminal is used for presenting the analysis result (specific image structure and corresponding area ratio) of the robust and ideal feature extraction network model obtained by the convolutional neural network model training unit to the input image to a doctor as a clinical diagnosis reference so as to improve the accuracy, the working efficiency and the working continuous state of the working personnel.
In this embodiment, the convolutional neural network constructing unit includes a feature extraction network and a classification network, and the feature extraction network extracts feature information of the image block, so that the subsequent classification network can correctly predict the relative position of the image block. In this embodiment, the classification network is composed of a fully connected layer and a LeakyReLU () activation function, and once training is completed, the classification network is discarded.
Feature extraction network this embodiment provides three models:
first, as shown in fig. 2, the feature extraction network is constructed by using the idea of modular concatenation, so that the width and depth of the network can be expanded as required, the lower layer of the feature extraction network is constructed by M convolution modules BLOCK-a, the upper layer of the feature extraction network is constructed by introducing N residual convolution modules BLOCK-B and P residual convolution modules BLOCK-C, and then a convolution layer and a tanh () activation function are connected with the upper layer. M is configurable, the value range is an integer between 3 and 6, the default value is 4, N is configurable, the value range is an integer between 1 and 3, and the default value is 2; p can be configured, the value range is a positive integer between 1 and 3, and the default value is 2.
The convolution module BLOCK-a is composed of a convolution layer and a LeakyReLU () activation function, as shown in FIG. 3; the residual convolution module BLOCK-B is formed by superimposing convolution kernels of 1x1 and 3x3 and layer skipping, as shown in fig. 4; the residual convolution module BLOCK-C is constructed by superimposing convolution kernels of 1x1, 1x3, 3x1 and layer jumps, as shown in fig. 5. Because the same layer has convolution kernels with different scales, sparse and non-sparse characteristics can be learned at the same time, and the jump layer (shortcuts) ensures that the network can consider deep and shallow network characteristics at the same time. The two characteristics in the network structure design increase the feature expression capability of the network.
Secondly, as shown in fig. 6, the feature extraction network is formed by using a modular splicing idea, so that the width and the depth of the network can be amplified as required, the lower layer of the feature extraction network is formed by K convolution modules BLOCK-D, the upper layer of the feature extraction network is introduced with Q self-supervision convolution modules BLOCK-F, and a convolution layer and a tanh () activation function are connected with the upper layer; k can be configured, the value range is an integer between 4 and 6, the default value is 5, Q can be configured, the value range is an integer between 1 and 3, and the default value is 1.
The convolution module BLOCK-D is composed of a convolution layer and a LeakyReLU () activation function, as shown in FIG. 7; the self-supervision convolution module BLOCK-F enables the same layer to have convolution kernels of various scales through a plurality of convolution kernels of 1x1 and 3x3 and a mean pooling layer which are mutually overlapped, so that sparse and non-sparse features can be simultaneously learned, and the feature expression capability of the network is increased, as shown in fig. 8.
Third, as shown in fig. 9, the feature extraction network includes a plurality of convolutional layers, and a BLOCK-G module is introduced in the middle of the plurality of convolutional layers. The feature extraction network extracts feature information of the image blocks so that a subsequent classification network can correctly predict the relative positions of the image blocks.
The BLOCK-G module includes convolution kernels of 1x1, 3x3, 5x5 and a maximum pooling layer, so that the same layer has convolution kernels of various scales, and the feature expression capability of the network is increased, as shown in fig. 10.
The convolution modules referred to in the present application all use conventional convolution layer modules, unless otherwise specified.
In order to improve the accuracy and robustness of the image anomaly detection model, the convolutional neural network unit further comprises a convolutional neural network model inspection unit; the convolutional neural network model checking unit comprises a model checking unit and a model testing unit, and the model checking unit is used for detecting the accuracy of the convolutional network model obtained by training; the model testing unit is used for detecting whether the convolutional network model obtained through training is over-fitted or not so as to screen out the network model with high robustness.
In this embodiment, when the analysis unit analyzes the specific image structure in the image to be analyzed by using the trained image anomaly detection model, the analysis unit may analyze the specific image structure by using the existing method, or may analyze the specific image structure by using the following method.
Image to be analyzed ItestAnd performing sliding window blocking, and dividing image blocks with the same size as the receiving field of the feature extraction network according to sliding step S pixels to obtain an image block sequence, wherein the size of the blocked image is W multiplied by W pixels, and S is more than or equal to 1 and less than or equal to W.
Carrying out self-adaptive segmentation on the image block subjected to sliding window partitioning, and distinguishing a target and a blank background in the image block; namely segmenting the foreground and the background, discarding the image blocks with the target ratio smaller than a threshold T1, and not performing subsequent processing; reserving image BLOCKs with a ratio greater than a threshold T1 to form a sequence of image BLOCKs { BLOCK }i,jAnd j and i are the counts of the image blocks on x and y coordinates respectively, and the counts and the y coordinates together form the number of the image blocks.
Image BLOCK sequence { BLOCKi,jInputting image blocks in the image block extraction network to obtain an image ItestThe abnormal feature map M of (1). The method specifically comprises the following steps: image BLOCK sequence { BLOCKi,jCalculating the abnormal value score abnormal thereof through a feature extraction networki,jAnd image BLOCK BLOCKi,jAbnormal value score of abnormali,jAs an initial outlier score, abnormal, for each pixel in the image blocki,j=minh∈S||f(p)-h||2Wherein the image BLOCK sequence { BLOCKi,jThe feature vector acquired by the input feature extraction network is f (BLOCK)i,j) H is any one of the feature vectors in the feature vector set S, | · | | non-calculation2Represents the L2 norm distance, then minh∈S||f(p)-h||2Then represents the image BLOCK BLOCKi,jIs spaced from the minimum L2 norm of any feature vector in the set S of feature vectors.
Then calculating the image I to be analyzedtestAn abnormal feature map M after the network is extracted through the features:
calculating an image I to be analyzedtestThe abnormal score value P of each pixel in the imagei,j
Figure BDA0002998437320000121
Image to be analyzed ItestAbnormal score value p of all pixels ini,jThe corresponding abnormal feature maps M, M, N are constructed to refer to the total number of image blocks in the x and y directions, respectively.
Performing threshold segmentation on the abnormal feature map M, and calculating the structural type of a specific image in the image I to be analyzed by using the segmented binary imagetestArea percentage of (1). The method specifically comprises the following steps:
performing threshold segmentation on the abnormal feature map M according to a threshold T2, and calculating the area percentage of a specific image structure
Figure BDA0002998437320000122
Wherein, AREAGCTIs an image I to be analyzedtestArea of region in (1), corresponding to sequence { BLOCKi,jSum of foreground AREA of each image block in }, AREASTRUCTIs an image I to be analyzedtestThe area of the specific image structure in the image is obtained by subtracting the sequence { BLOCK } from the foreground area after the threshold segmentation of the corresponding abnormal feature map Mi,jThe sum of the background areas of each image block in the pixel.
The application also provides a method for constructing the image analysis system, which specifically comprises the following steps: constructing the image analysis system, acquiring an original image by an image database construction unit, labeling different image structures in the original image, classifying and sorting the labeled image data, and dividing a training set, a check set and a test set; training the image anomaly detection model by adopting a training set to obtain an ideal image anomaly detection model, and verifying the ideal image anomaly detection model by adopting a verification set to detect the accuracy of the ideal image anomaly detection model; and testing the ideal image anomaly detection model by adopting a test set to detect the robustness of the model.
If the difference between the accuracy of the image anomaly detection model on the test set and the accuracy in the training of the check set exceeds a preset value, the model is over-fitted, the model returns to the convolutional neural network training unit, the network structure or the parameters are adjusted to perform retraining again, so that the image anomaly detection model with the difference between the accuracy on the test set and the accuracy in the training of the check set within the preset value is obtained, and at the moment, the robustness of the image anomaly detection model is high.
The training flow of the training set to the image anomaly detection model is as follows:
b1: and training a feature extraction network of the image anomaly detection model by adopting the images in the training set.
In this embodiment, two training procedures are provided as follows:
the first method comprises the following steps:
b11: for each image in the training set, as shown in fig. 11, an image block p is arbitrarily selected from eight neighborhoods of a 3 × 3 grid, the scale of p is the same as the receiving field of the feature extraction network, then random dithering is performed on the center of the image block p to obtain an image block p1, the cross entropy of the image block p and p1 is calculated to be used as a sub-item Loss function Loss _1,
Figure BDA0002998437320000131
wherein the image block p1The true relative position with respect to p is y {0,1, …,7}, yiReferring to the number of image blocks of the 8 relative positions of the category i in y {0,1, …,7} in the training set; classifier CφIs trained to correctly predict image block p1Relative to the image block p, i.e. y ═ Cφ(fθ(p),fθ(p1) Representation of an image block p predicted by a classifier1Relative position to image block p, aiIs the confidence of class i calculated by the classifier, and N is the total number of samples in the training set.
For the image block p, as shown in fig. 12, in the four neighborhoods of the 5 × 5 grid, an image block p in the same row or column but not adjacent to the image block p is randomly selected2,p2The scale is the same as the acceptance field of the feature extraction network, the cross entropy of the image blocks p and p2 is calculated as a sub-term Loss function Loss _2,
Figure BDA0002998437320000141
wherein the image block p2The true relative position with respect to p is y {0,1,2,3}, yiReferring to the number of image blocks of the 4 relative positions of the category i in the training set, namely y {0,1,2,3 }; classifier CφIs trained to correctly predict image block p2And imagesRelative position of block p, i.e. y ═ Cφ(fθ(p),fθ(p2) Representation of an image block p predicted by a classifier2Relative position to image block p, biIs the confidence level for class i calculated by the classifier.
For an image block p, as shown in fig. 13, for four adjacent intersection areas of p, 2 to 4 of the image blocks p3, p4, p5, and p6 are obtained, that is, for four edge points (upper left corner, upper right corner, lower left corner, and lower right corner) of the image block p as new image block centers, four new image blocks p3, p4, p5, and p6 with the same scale as the image block p can be obtained, L2 norm distances of selected image blocks among p and p3, p4, p5, and p6 are calculated, and averaged to be used as a sub-term Loss function Loss _3,
Figure BDA0002998437320000142
||fθ(p)-fθ(p2+i)||2the L2 norm distance between the image block p and the selected image block in p3, p4, p5 and p6 is pointed out.
B12: calculating Loss function Loss of network model as lambda1*Loss_1+λ2*Loss_2+Loss_3,λ1、λ2And performing back propagation by using an Adam optimizer to realize network weight iteration and optimization of the feature extraction network model, wherein the weight values in the loss function are all larger than 0.
B13: and C, repeatedly executing the steps B11-B12 until the number of turns is specified, and selecting and storing the optimal weight of the feature extraction network and the classification network according to the loss function of each turn of training.
And the second method comprises the following steps:
step 1: for each image in the training set, randomly selecting an image block p7 in the eight neighborhood of the 3x3 grid of the image block p, calculating the cross entropy of the image blocks p and p7 as a sub-item Loss function Loss _4,
Figure BDA0002998437320000151
wherein the image block p7The true relative position with respect to p is y {0,1, …,7}, yiIn the training set, the class i is relative to the class i in the 8 pairs of y {0,1, …,7}The number of image blocks of a location; classifier CφIs trained to correctly predict image block p7Relative to the image block p, i.e.
Figure BDA0002998437320000153
Representing a predicted image block p by a classifier7Relative position to image block p, ciIs the confidence of class i calculated by the classifier, and N is the total number of samples in the training set.
Randomly taking an image block p8 which is in the same row or column but is not adjacent to the image block p at intervals of four neighborhoods of a 5X5 network aiming at the image block p, calculating the cross entropy of p and p8 as a sub-term Loss function Loss _5,
Figure BDA0002998437320000152
wherein the image block p8The true relative position with respect to p is y {0,1,2,3}, yiReferring to the number of image blocks of the 4 relative positions of the category i in the training set, namely y {0,1,2,3 }; classifier CφIs trained to correctly predict image block p8With respect to the relative position of the image block p,
Figure BDA0002998437320000154
representing a predicted image block p by a classifier8Relative position to image block p, diIs the confidence level for class i calculated by the classifier.
Step 2: and calculating a Loss function Loss of the network model, wherein the Loss is lambda Loss _4+ Loss _5, the lambda is a weight value in the Loss function and is larger than 0, and performing back propagation by using an Adam optimizer to realize network weight iteration and optimization of the feature extraction network model.
And step 3: and (3) repeatedly executing the steps 1-2 to designate the number of rounds, and selecting and storing the optimal weights of the feature extraction network and the classification network according to the loss function of each round of training.
It should be noted that if the number of image data in the training set is too small, the minimum number of network layers can be set in a configurable range for the network structure, so as to avoid overfitting when the training data is insufficient; if training is carried out on the conventional convolutional neural network model, the model accuracy cannot be increased, more network layers can be set in a configurable range according to the network structure, namely, the fitting capability of the model is improved by increasing the depth of the convolutional model.
Then the next step is performed:
b2: using the trained feature extraction network to obtain and store a feature set S, S ← S { (f) } corresponding to the training setθ(p) }, i.e.: randomly extracting an image block aiming at each image in the training set, wherein the image block has the same receiving field as the feature extraction network, and acquiring a feature vector f of the image block through the feature extraction network obtained by trainingθ(p) the feature vectors as a whole constitute a feature vector set S.
B3: saving trained feature extraction network weights fθAnd training a feature set S corresponding to the image set to finish the training of the image anomaly detection model.
After the image anomaly detection model is trained, an ideal image anomaly detection model is obtained, and the ideal image anomaly detection model can be verified by adopting a verification set to detect the accuracy of the ideal image anomaly detection model; testing an ideal image anomaly detection model by adopting a test set to detect the robustness of the ideal image anomaly detection model; if the difference between the accuracy of the image anomaly detection model on the test set and the accuracy in the training of the check set exceeds a preset value, the model is over-fitted, the network structure or the parameters are adjusted to train again to obtain the image anomaly detection model with the difference between the accuracy on the test set and the accuracy in the training of the check set within the preset value, and at the moment, the robustness of the image anomaly detection model is high. The accuracy here is: the method comprises the steps of conducting sliding window blocking on images in a check set or a test set according to step length pixels to obtain an image block sequence (the image block sequence is the same as a receiving field of a feature extraction network), then obtaining an abnormal feature map of the images through the feature extraction network, conducting threshold segmentation on the abnormal feature map, calculating the area percentage of a specific image structure type in the images by utilizing the segmented binary images, and comparing the percentage with an artificial labeling result to obtain accuracy. And if the accuracy of the images in the verification set is within the acceptable range, the image anomaly detection model obtained by training is considered as an ideal model.
The image analysis system and the construction method can be used for analyzing the pathological image, but not limited to analyzing the pathological image, so as to obtain the following image structure of tumor cells in the pathological image: the intelligent, efficient and quantitative analysis of the follicular structure, the island structure, the funicular structure, the ribbon structure, the diffuse structure and the like is taken as an example, 8 pathological doctors with diagnosis experience of ovarian granulosa cytoma for more than 5 years are selected, 30 pathological section images of the ovarian granulosa cytoma are provided for each person respectively, the tumor cell image structure in the pathological section images is analyzed, then the accuracy and the average time are calculated, the diagnosis state of the doctors is counted, and compared with the scheme provided by the invention, the results are shown in the following table 1.
TABLE 1 comparison of the results of the image analysis of ovarian granulosa cell tumors in pathological sections
Figure BDA0002998437320000171
As can be seen from table 1, the accuracy of the analysis of the tumor cell image structure (follicular structure, island structure, funicular structure, ribbon structure, diffuse structure, etc.) in the pathological section by using the scheme provided by the present invention is higher than that of a professional pathologist, and a quantitative conclusion can be obtained (the pathologist can only obtain a subjective qualitative or semi-quantitative conclusion by visual analysis). In addition, the method of the invention has shorter analysis time and long work duration.
In the description herein, references to the description of the term "one embodiment," "some embodiments," "an example," "a specific example," or "some examples," etc., mean that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the invention. In this specification, the schematic representations of the terms used above do not necessarily refer to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples.
While embodiments of the invention have been shown and described, it will be understood by those of ordinary skill in the art that: various changes, modifications, substitutions and alterations can be made to the embodiments without departing from the principles and spirit of the invention, the scope of which is defined by the claims and their equivalents.

Claims (9)

1. An image analysis system is characterized by comprising an image database construction unit, a convolutional neural network unit and an analysis unit;
the image database construction unit comprises an image data acquisition unit for acquiring input image data, an image data labeling unit for labeling different image structures in each input image data, and an image database construction unit for classifying and sorting the labeled image data provided by the image data labeling unit;
the convolutional neural network unit comprises a convolutional neural network model construction unit and a convolutional neural network model training unit, and the convolutional neural network model construction unit is used for constructing an image anomaly detection model; the convolutional neural network model training unit trains the image anomaly detection model, inputs the model into a training image set, loss function weights, a feature extraction network and a classification network, and outputs a feature set S comprising the training set, a trained feature extraction network and weights f thereofθ
And the analysis unit analyzes the specific image structure in the image to be analyzed by using the trained image anomaly detection model.
2. The image analysis system according to claim 1, wherein the convolutional neural network construction unit includes a feature extraction network and a classification network;
the feature extraction network comprises a lower layer formed by M convolution modules BLOCK-A, a higher layer formed by N residual convolution modules BLOCK-B, P residual convolution modules BLOCK-C, a convolution layer connected with the higher layer and a tanh () activation function;
the convolution module BLOCK-A consists of a convolution layer and a LeakyReLU () activation function;
the residual convolution module BLOCK-B is composed of superposed convolution kernels of 1x1 and 3x3 and a skip layer;
the residual convolution module BLOCK-C is composed of a convolution kernel and a layer jump, wherein the convolution kernel is superposed with 1x1, 1x3 and 3x 1;
the classification network consists of a full connectivity layer and a LeakyReLU () activation function.
3. The image analysis system according to claim 1, wherein the convolutional neural network construction unit includes a feature extraction network and a classification network;
the feature extraction network comprises a lower layer formed by K convolution modules BLOCK-D, a higher layer formed by Q self-supervision convolution modules BLOCK-F, a convolution layer connected with the higher layer and a tanh () activation function;
the convolution module BLOCK-D consists of a convolution layer and a LeakyReLU () activation function;
the self-supervision convolution module BLOCK-F module comprises a plurality of 1x1 and 3x3 convolution kernels and a mean pooling layer which are mutually overlapped;
the classification network consists of a full connectivity layer and a LeakyReLU () activation function.
4. The image analysis system according to claim 1, wherein the convolutional neural network construction unit includes a feature extraction network and a classification network;
the feature extraction network comprises a plurality of convolution layers, and a BLOCK-G module is introduced in the middle of the convolution layers;
the BLOCK-G module comprises a plurality of superimposed 1x1, 3x3, 5x5 convolution kernels and a max pooling layer;
the classification network consists of a full connectivity layer and a LeakyReLU () activation function.
5. The image analysis system of claim 1, wherein the convolutional neural network unit further comprises a convolutional neural network model verification unit;
the convolutional neural network model checking unit comprises a model checking unit and a model testing unit, and the model checking unit is used for detecting the accuracy of the convolutional network model obtained by training; the model testing unit is used for detecting whether the convolutional network model obtained through training is over-fitted or not so as to screen out the network model with high robustness.
6. An image analysis system construction method is characterized in that the image analysis system according to any one of claims 1 to 5 is constructed, an image database construction unit obtains an original image and labels different image structures in the original image, the labeled image data is classified and sorted, and a training set, a check set and a test set are divided; training the image anomaly detection model by adopting a training set to obtain an ideal image anomaly detection model, and verifying the ideal image anomaly detection model by adopting a verification set to detect the accuracy of the ideal image anomaly detection model; testing an ideal image anomaly detection model by adopting a test set to detect the robustness of the ideal image anomaly detection model;
if the difference between the accuracy of the image anomaly detection model on the test set and the accuracy in the training of the check set exceeds a preset value, the model is over-fitted, the model returns to the convolutional neural network training unit, the network structure or the parameters are adjusted to perform retraining again, so that the image anomaly detection model with the difference between the accuracy on the test set and the accuracy in the training of the check set within the preset value is obtained, and at the moment, the robustness of the image anomaly detection model is high.
7. The image analysis system construction method according to claim 6, wherein the training flow of the training set to the image anomaly detection model is as follows:
b1: training a feature extraction network of an image anomaly detection model by using images in a training set;
b2: using the trained feature extraction network to obtain and store a feature set S, S ← S { (f) } corresponding to the training setθ(p) }, i.e.: randomly extracting an image block for each image in the training set, wherein the image block is the same as the receiving field of the feature extraction network, and acquiring the image block by the feature extraction network obtained by trainingThe feature vector f of the image blockθ(p) the feature vectors as a whole constitute a feature vector set S;
b3: saving trained feature extraction network weights fθAnd a feature vector set S corresponding to the training set.
8. The image analysis system construction method according to claim 7, wherein the step B1 includes the steps of:
b11: for each image in the training set, randomly selecting an image block p in an eight-neighborhood of a 3x3 grid of the image, wherein the scale of the image block p is the same as the receiving field of the feature extraction network, then randomly dithering the center of the image block p to obtain an image block p1, calculating the cross entropy of the image block p and the image block p1 as a sub-term Loss function Loss _1,
Figure FDA0002998437310000041
wherein the image block p1The true relative position with respect to p is y {0,1, …,7}, yiReferring to the number of image blocks of the 8 relative positions of the category i in y {0,1, …,7} in the training set; classifier CφIs trained to correctly predict image block p1Relative to the image block p, i.e. y ═ Cφ(fθ(p),fθ(p1)),aiThe confidence of the category i calculated by the classifier, and N is the total number of samples in the training set;
for an image block p, randomly selecting an image block p which is in the same row or column but not adjacent to the image block p in the four neighborhoods of the 5 × 5 grid2,p2The scale is the same as the acceptance field of the feature extraction network, the cross entropy of the image blocks p and p2 is calculated as a sub-term Loss function Loss _2,
Figure FDA0002998437310000042
wherein the image block p2The true relative position with respect to p is y {0,1,2,3}, yiReferring to the number of image blocks of the 4 relative positions of the category i in the training set, namely y {0,1,2,3 }; classifier CφIs trained to correctly predict image block p2Phase with image block pTo position, i.e. y ═ Cφ(fθ(p),fθ(p2)),biIs the confidence of class i calculated by the classifier;
aiming at an image block p, acquiring 2-4 image blocks p3, p4, p5 and p6 of four adjacent intersection areas of p, calculating L2 norm distances between p and selected image blocks p3, p4, p5 and p6, and averaging the distances to be used as a subitem Loss function Loss _3,
Figure FDA0002998437310000043
||fθ(p)-fθ(p2+i)||2the L2 norm distance between the image block p and a selected image block in p3, p4, p5 and p 6;
b12: calculating Loss function Loss of network model as lambda1*Loss_1+λ2*Loss_2+Loss_3,λ1、λ2The weighted value in the loss function is larger than 0, and backward propagation is carried out by using an Adam optimizer, so that network weight iteration and optimization of the feature extraction network model are realized;
b13: and C, repeatedly executing the steps B11-B12 until the number of turns is specified, and selecting and storing the optimal weight of the feature extraction network and the classification network according to the loss function of each turn of training.
9. The image analysis system construction method according to claim 7, wherein the step B1 includes the steps of:
step 1: for each image in the training set, randomly selecting an image block p7 in the eight neighborhood of the 3x3 grid of the image block p, calculating the cross entropy of the image blocks p and p7 as a sub-item Loss function Loss _4,
Figure FDA0002998437310000051
wherein the image block p7The true relative position with respect to p is y {0,1, …,7}, yiReferring to the number of image blocks of the 8 relative positions of the category i in y {0,1, …,7} in the training set; classifier CφIs trained to correctly predict image block p7Relative bit to image block pPut into, i.e.
Figure FDA0002998437310000054
ciIs the probability value of the category i calculated by the classifier, and N is the total number of samples in the training set;
for an image block p, randomly taking an image block p8 which is in the same row or column but not adjacent to the image block p in the four neighborhoods of the 5X5 network, calculating the cross entropy of p and p8 as a sub-term Loss function Loss _5,
Figure FDA0002998437310000052
wherein the image block p8The true relative position with respect to p is y {0,1,2,3}, yiReferring to the number of image blocks of the 4 relative positions of the category i in the training set, namely y {0,1,2,3 }; classifier CφIs trained to correctly predict image block p8Relative to the image block p, i.e.
Figure FDA0002998437310000053
diIs the probability value of the category i calculated by the classifier;
step 2: calculating a Loss function Loss of the network model, wherein lambda is a weight value in the Loss function and is larger than 0, and performing back propagation by using an Adam optimizer to realize network weight iteration and optimization of the feature extraction network model;
and step 3: and (3) repeatedly executing the steps 1-2 to designate the number of rounds, and selecting and storing the optimal weights of the feature extraction network and the classification network according to the loss function of each round of training.
CN202110338180.4A 2021-03-30 2021-03-30 Image analysis system and construction method thereof Active CN113096079B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110338180.4A CN113096079B (en) 2021-03-30 2021-03-30 Image analysis system and construction method thereof

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110338180.4A CN113096079B (en) 2021-03-30 2021-03-30 Image analysis system and construction method thereof

Publications (2)

Publication Number Publication Date
CN113096079A true CN113096079A (en) 2021-07-09
CN113096079B CN113096079B (en) 2023-12-29

Family

ID=76671154

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110338180.4A Active CN113096079B (en) 2021-03-30 2021-03-30 Image analysis system and construction method thereof

Country Status (1)

Country Link
CN (1) CN113096079B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113691940A (en) * 2021-08-13 2021-11-23 天津大学 Incremental intelligent indoor positioning method based on CSI image

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108922602A (en) * 2018-05-28 2018-11-30 中山大学附属第六医院 The same period new chemoradiation therapy curative effect evaluation system and method before rectal cancer based on big data analysis MRI image
WO2019099226A1 (en) * 2017-11-14 2019-05-23 Google Llc Weakly-supervised action localization by sparse temporal pooling network
WO2020014477A1 (en) * 2018-07-11 2020-01-16 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for image analysis with deep learning to predict breast cancer classes
CN110727097A (en) * 2019-12-19 2020-01-24 上海兰脉信息科技有限公司 Pathological microscopic image real-time acquisition and analysis system, method, device and medium
CN111178245A (en) * 2019-12-27 2020-05-19 深圳佑驾创新科技有限公司 Lane line detection method, lane line detection device, computer device, and storage medium
CN111666998A (en) * 2020-06-03 2020-09-15 电子科技大学 Endoscope intelligent intubation decision-making method based on target point detection
CN111985536A (en) * 2020-07-17 2020-11-24 万达信息股份有限公司 Gastroscope pathological image classification method based on weak supervised learning
CN112364565A (en) * 2020-11-12 2021-02-12 河北工业大学 Fault arc protection method based on AlexNet
CN112446388A (en) * 2020-12-05 2021-03-05 天津职业技术师范大学(中国职业培训指导教师进修中心) Multi-category vegetable seedling identification method and system based on lightweight two-stage detection model

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2019099226A1 (en) * 2017-11-14 2019-05-23 Google Llc Weakly-supervised action localization by sparse temporal pooling network
CN108922602A (en) * 2018-05-28 2018-11-30 中山大学附属第六医院 The same period new chemoradiation therapy curative effect evaluation system and method before rectal cancer based on big data analysis MRI image
WO2020014477A1 (en) * 2018-07-11 2020-01-16 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for image analysis with deep learning to predict breast cancer classes
CN110727097A (en) * 2019-12-19 2020-01-24 上海兰脉信息科技有限公司 Pathological microscopic image real-time acquisition and analysis system, method, device and medium
CN111178245A (en) * 2019-12-27 2020-05-19 深圳佑驾创新科技有限公司 Lane line detection method, lane line detection device, computer device, and storage medium
CN111666998A (en) * 2020-06-03 2020-09-15 电子科技大学 Endoscope intelligent intubation decision-making method based on target point detection
CN111985536A (en) * 2020-07-17 2020-11-24 万达信息股份有限公司 Gastroscope pathological image classification method based on weak supervised learning
CN112364565A (en) * 2020-11-12 2021-02-12 河北工业大学 Fault arc protection method based on AlexNet
CN112446388A (en) * 2020-12-05 2021-03-05 天津职业技术师范大学(中国职业培训指导教师进修中心) Multi-category vegetable seedling identification method and system based on lightweight two-stage detection model

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
M. LI等: "ComNet: Combinational Neural Network for Object Detection in UAV-Borne Thermal Images", 《IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING》, vol. 59, no. 8, XP011867608, DOI: 10.1109/TGRS.2020.3029945 *
廖欣等: "基于深度卷积神经网络的宫颈细胞病理智能辅助诊断方法", 《液晶与显示》, vol. 33, no. 6 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113691940A (en) * 2021-08-13 2021-11-23 天津大学 Incremental intelligent indoor positioning method based on CSI image
CN113691940B (en) * 2021-08-13 2022-09-27 天津大学 Incremental intelligent indoor positioning method based on CSI image

Also Published As

Publication number Publication date
CN113096079B (en) 2023-12-29

Similar Documents

Publication Publication Date Title
EP3961484B1 (en) Medical image segmentation method and device, electronic device and storage medium
CN110287932B (en) Road blocking information extraction method based on deep learning image semantic segmentation
CN106682633B (en) The classifying identification method of stool examination image visible component based on machine vision
CN111931811B (en) Calculation method based on super-pixel image similarity
CN102722712B (en) Multiple-scale high-resolution image object detection method based on continuity
CN109614985A (en) A kind of object detection method based on intensive connection features pyramid network
CN107945153A (en) A kind of road surface crack detection method based on deep learning
CN107977671A (en) A kind of tongue picture sorting technique based on multitask convolutional neural networks
CN109389129A (en) A kind of image processing method, electronic equipment and storage medium
Morris A pyramid CNN for dense-leaves segmentation
CN106529499A (en) Fourier descriptor and gait energy image fusion feature-based gait identification method
Pan et al. Cell detection in pathology and microscopy images with multi-scale fully convolutional neural networks
CN105260738A (en) Method and system for detecting change of high-resolution remote sensing image based on active learning
CN104484886B (en) A kind of dividing method and device of MR images
CN106408030A (en) SAR image classification method based on middle lamella semantic attribute and convolution neural network
CN106778687A (en) Method for viewing points detecting based on local evaluation and global optimization
CN106096613A (en) Image multi-target detection method and device based on corner feature
CN101667245A (en) Human face detection method by cascading novel detection classifiers based on support vectors
CN112257665A (en) Image content recognition method, image recognition model training method, and medium
CN110349167A (en) A kind of image instance dividing method and device
WO2022152009A1 (en) Target detection method and apparatus, and device and storage medium
CN114600155A (en) Weakly supervised multitask learning for cell detection and segmentation
CN115546605A (en) Training method and device based on image labeling and segmentation model
CN110264454A (en) Cervical cancer tissues pathological image diagnostic method based on more hidden layer condition random fields
Yadav et al. An improved deep learning-based optimal object detection system from images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant