CN111369540A - Plant leaf disease identification method based on mask convolutional neural network - Google Patents
Plant leaf disease identification method based on mask convolutional neural network Download PDFInfo
- Publication number
- CN111369540A CN111369540A CN202010150980.9A CN202010150980A CN111369540A CN 111369540 A CN111369540 A CN 111369540A CN 202010150980 A CN202010150980 A CN 202010150980A CN 111369540 A CN111369540 A CN 111369540A
- Authority
- CN
- China
- Prior art keywords
- mask
- target
- network
- image
- training
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 201000010099 disease Diseases 0.000 title claims abstract description 73
- 208000037265 diseases, disorders, signs and symptoms Diseases 0.000 title claims abstract description 73
- 238000000034 method Methods 0.000 title claims abstract description 37
- 238000013527 convolutional neural network Methods 0.000 title claims description 40
- 238000012549 training Methods 0.000 claims abstract description 47
- 238000012360 testing method Methods 0.000 claims abstract description 31
- 238000001514 detection method Methods 0.000 claims abstract description 28
- 230000011218 segmentation Effects 0.000 claims abstract description 16
- 238000012216 screening Methods 0.000 claims abstract description 15
- 238000013528 artificial neural network Methods 0.000 claims abstract description 4
- 230000002708 enhancing effect Effects 0.000 claims abstract description 4
- 239000013598 vector Substances 0.000 claims description 19
- 238000010586 diagram Methods 0.000 claims description 9
- 238000012545 processing Methods 0.000 claims description 7
- 238000013507 mapping Methods 0.000 claims description 6
- 230000003044 adaptive effect Effects 0.000 claims description 4
- 230000004913 activation Effects 0.000 claims description 2
- 238000013519 translation Methods 0.000 claims description 2
- 230000003190 augmentative effect Effects 0.000 claims 2
- 238000000605 extraction Methods 0.000 claims 1
- 238000000844 transformation Methods 0.000 claims 1
- 241000196324 Embryophyta Species 0.000 description 28
- 238000002474 experimental method Methods 0.000 description 5
- 240000008067 Cucumis sativus Species 0.000 description 4
- 238000013135 deep learning Methods 0.000 description 4
- 235000010799 Cucumis sativus var sativus Nutrition 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 238000004088 simulation Methods 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000006378 damage Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 241001270131 Agaricus moelleri Species 0.000 description 1
- 235000009849 Cucumis sativus Nutrition 0.000 description 1
- 241000221785 Erysiphales Species 0.000 description 1
- 206010027146 Melanoderma Diseases 0.000 description 1
- 241000233679 Peronosporaceae Species 0.000 description 1
- 241000700605 Viruses Species 0.000 description 1
- 230000001580 bacterial effect Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 235000013399 edible fruits Nutrition 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 230000005541 medical transmission Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 239000002689 soil Substances 0.000 description 1
- 238000011282 treatment Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/22—Matching criteria, e.g. proximity measures
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/73—Deblurring; Sharpening
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/136—Segmentation; Edge detection involving thresholding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/462—Salient features, e.g. scale invariant feature transforms [SIFT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/30—Subject of image; Context of image processing
- G06T2207/30181—Earth observation
- G06T2207/30188—Vegetation; Agriculture
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A40/00—Adaptation technologies in agriculture, forestry, livestock or agroalimentary production
- Y02A40/10—Adaptation technologies in agriculture, forestry, livestock or agroalimentary production in agriculture
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Bioinformatics & Computational Biology (AREA)
- General Engineering & Computer Science (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Quality & Reliability (AREA)
- Multimedia (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a plant leaf disease identification method based on a mask convolution neural network, which mainly solves the problem of low accuracy rate of identifying plant leaf diseases in the prior art. The scheme is as follows: enhancing and expanding the original data set to obtain a training set and a test set; performing semantic segmentation on the training set and the test set to obtain a corresponding mask set; adding a disease characteristic screening module between the full convolution layer and the mask branch of the model, inputting the training set and the mask set into a network for training, and obtaining the results of target classification and target detection; taking a characteristic graph belonging to disease leaves in a target classification result as the input of a mask branch, and obtaining a trained model after multiple iterations; and inputting the test set into the model, performing target classification and target detection on the leaves, and segmenting the leaves belonging to the disease category. The method improves the accuracy of leaf disease identification on the basis of the traditional method, and can be used for identifying and segmenting plant disease leaves in agricultural planting.
Description
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a plant leaf disease identification method which can be used for segmenting and identifying plant disease leaves in agricultural planting.
Background
In modern wisdom agriculture, crop diseases are a great threat to grain safety, and plant diseases cause serious damage to crops by significantly reducing yield. Early blight is a typical disease that can severely reduce yield. Similarly, in humid climates, late blight is another very devastating disease that affects the leaves, stems and fruits of plants. Protection of plants from disease is critical to ensure the quality and quantity of crops. Crop protection should begin with the early detection of disease so that appropriate treatments can be selected at the right time to prevent disease transmission. The method for identifying the disease types of the greenhouse plant diseases mainly comprises various diseases such as bacterial spots, early blight, late blight, leaf mold, black spot and the like. However, in practice, the number of diseases is large, and the diseases are similar in performance on leaves, so that it is difficult to accurately determine the types of the diseases.
At present, the disease research on plant leaves mainly comprises detecting and classifying the plant leaves by using an image processing or deep learning method. In the Plant disease control, a foreign Plant Village group provides a method for detecting and classifying diseases through Plant leaves by utilizing deep learning in 2016, mainly classifies specific partial Plant diseases through a smart phone under a simple background, and mainly identifies colors, gray scales and segmented pictures through the deep learning method under data sets with different proportions and different networks, but the method can only classify the Plant disease leaves and cannot segment the Plant disease leaves and the positions of the diseases.
The same year Mads Dyrmann1 et al proposed the use of convolutional neural networks to classify plant species in color images. The network is built from scratch, trained and tested. These images are from six different local datasets, with data acquisition at different growth stages, in terms of illumination, resolution and soil type, but with low accuracy and without detection and classification of plant diseases.
Liuna et al in 2018 use an image processing technology and an artificial neural network technology to realize disease detection and disease degree classification of cucumber leaves, and mainly perform experimental research on cucumber downy mildew, powdery mildew and virus diseases with high morbidity and serious harm, but because the disease types of cucumbers are identified less, the number of training samples is less, and the identified disease types are less, the test accuracy is lower and overfitting is easily generated.
The framework of Mask convolutional neural network Mask R-CNN is provided by Heterokamm in 2017, and is characterized in that semantic segmentation is carried out on two branches of FasterR-CNN, namely, on classification and coordinate regression, one branch is added, image features are extracted mainly through a residual error network ResNet101/50 or a pyramid network FPN serving as a main network, the foreground and the background of a target region are obtained through a region recommendation network RPN, classification and example segmentation results are obtained through full convolutional layers on the obtained target region and the image features, and then the semantic segmentation results are obtained through convolutional network identification. The network is mainly used for target detection of a COCO data set, and is not used for the field of plant disease identification at present.
In summary, the current research on plant diseases mainly includes classification and identification of single plant disease types similar to cucumber, wherein the included samples include fewer disease types and samples, and the identification accuracy is low. The existing method for classifying the plant disease leaves through deep learning cannot separate the plant disease leaves and the positions of diseases, and the recognition rate is low.
Disclosure of Invention
The invention aims to provide a plant leaf disease identification method based on mask convolution neural network MaskR-CNN aiming at the defects in the prior art, so as to segment plant disease leaves and the positions of diseases, and improve the identification accuracy and efficiency.
In order to achieve the purpose, the technical scheme of the invention comprises the following steps:
(1) sequentially enhancing, expanding and semantically segmenting an original data set to obtain a training and testing image set and a mask set;
(2) optimizing a Mask R-CNN network: a disease characteristic screening module is added between the full convolution layer and the Mask branch after the Mask R-CNN network ROI Align;
(3) training the optimized Mask R-CNN network:
(3a) setting main network parameters:
selecting a main network from the ResNet50 and ResNet101 residual error networks;
setting the size of the epoch and the step number of each epoch training;
setting a receiving threshold T of a diseased leaf in a diseased feature screening module0The other parameters are default values of Mask R-CNN;
(3b) according to the known classification error LclsDetecting error LboxAnd a segmentation error LmaskAnd determining the optimized MaskR-CNN network loss function as follows: l isloss=Lcls+Lbox+Lmask;
(3c) Inputting the training image set and the training Mask set into an optimized Mask R-CNN network for training to obtain a trained model;
(4) and inputting the test image into the trained model for testing.
Compared with the prior art, the invention has the following advantages:
firstly, compared with the GoogleNet and VGG methods, the method disclosed by the invention has the advantage that the identification accuracy of the damaged leaves is improved when the plant leaf diseases are identified.
Secondly, the Mask image of the target area is generated through the Mask branch of the Mask R-CNN network model, so that the plant disease leaves and the disease positions thereof can be accurately extracted.
Thirdly, because the disease characteristic screening module is added in the network structure of the Mask R-CNN, the invention only aims at unhealthy leaves to carry out the training and testing of Mask branches, reduces the burden of the Mask branches and improves the identification efficiency of the network on the premise of ensuring that the accuracy of the Mask R-CNN network is not changed.
Fourthly, in the invention, because the samples in the data set are increased by adopting image transformation and the data set is enhanced by adopting an adaptive contrast enhancement algorithm, the blurred image in the data set is improved.
Drawings
FIG. 1 is a general flow diagram of an implementation of the present invention;
FIG. 2 is a schematic diagram of the overall structure of an optimized Mask R-CNN network in the present invention;
FIG. 3 is a block diagram of a training sub-process for an optimized Mask R-CNN network in the present invention;
FIG. 4 is a training image and a training binary mask image acquired in the present invention;
FIG. 5 is a test image of healthy and diseased leaves taken in the present invention;
FIG. 6 is a result image of healthy leaves identified by simulation of the present invention;
FIG. 7 is a result image of diseased leaves identified by simulation in accordance with the present invention.
Detailed Description
The following detailed description of the embodiments and effects of the present invention will be made with reference to the accompanying drawings:
the application environment of the embodiment is an agricultural planting scene, and aims to detect and identify vegetation with diseases in agricultural planting and provide disease information for planting personnel more accurately.
Referring to fig. 1, the implementation steps of this example are as follows:
step 1, image enhanced data set D1。
(1.1) plant disease leaf Dataset D was downloaded from public project PlantVillage-Dataset0Using adaptive contrast enhancement algorithm to pair D0Performing image enhancement to improve the blurred image in the database to obtain an image-enhanced database D1:
(1.1a) acquiring the low frequency part m of the image x (i, j)x(i, j) and a high frequency part hx(i, j), obtaining the low frequency part of the image by means filtering:
hx(i,j)=x(i,j)-mx(i,j)
wherein, (2n +1)2(ii) represents the window size with (i, j) as the image center point coordinate;
(1.1b) multiplying the high frequency part in the image by the gain value G (I, j) of the part to obtain an amplified high frequency part I (I, j):
I(i,j)=G(i,j)hx(i,j)
wherein G (I, j) may take a constant C greater than 1 to obtain an amplified high-frequency portion Ic(i, j) is:
Ic(i,j)=Chx(i,j)
in this example the gain value G (i, j) is taken to be equal to the local mean square error σx(i, j) is inversely proportional to the change valueWherein D is a constant, and the local mean square error of the image is:
obtaining an amplified high-frequency part Iσ(i, j) is:
(1.1c) recombining the high frequency part and the low frequency part to obtain an enhanced image f (i, j):
f(i,j)=mx(i,j)+Iσ(i,j);
step 2, obtaining a training image set D3And testing the image set D4And a training mask set D5And test mask set D6。
(2.1) enhancing the image data set D by using a semantic segmentation labeling tool labelme1Respectively plotting the image targets to generate the mask of the targets, and obtaining a mask set D containing mask information and label information2The resulting training binary mask image is shown in fig. 4 b.
(2.2) Using image transformation on the enhanced dataset D1Sum mask set D2Sequentially translating, rotating and overturning to increase the sample data volume and obtain a data set D after sample expansion3Sum mask set D4;
(2.2) expanding the data set D3Sum mask set D4Dividing the training image into training image sets D according to the proportion of 8:25And testing the image set D6And a training mask set D7And test mask set D8The obtained training set image is shown in fig. 4a, and the test set image is shown in fig. 5a and 5 b;
and step 3, continuously optimizing the Mask R-CNN network structure.
The existing Mask R-CNN network comprises a main network, a regional recommendation generation network RPN, a full convolution layer and a Mask branch, namely the full convolution layer and a full connection layer, in the embodiment, a disease characteristic screening module is added between the full convolution layer and the Mask branch to obtain an optimized Mask R-CNN network structure, which is shown in figure 2.
Referring to fig. 2, the Mask R-CNN network structure after optimization in this example is: the main network → the area recommendation generation network RPN → the full convolution layer → the disease feature screening module → the full convolution layer → the full connection layer.
The disease characteristic screening moduleThe block is to judge the confidence coefficient T of the fault blade output by the full convolution layer1Receiving threshold T of damaged leaf given by network initialization parameter0The characteristic maps of the diseased leaves in batch processing are screened out, and screening results are input to mask branches, namely, full convolution layers and full connection for processing.
Step 4, training the optimized Mask R-CNN network to obtain a trained network model:
referring to fig. 3, the specific implementation of this step is as follows:
(4.1) setting main network parameters:
selecting a main network from the ResNet50 and ResNet101 residual error networks, wherein ResNet101 is selected as the main network in the example;
setting the number of labels to be 11 according to the category of the images in the database, wherein the number of labels comprises 1 background label and 10 image labels;
setting the iteration times epoch of all samples to be 100, the iteration times of each epoch to be 100, the learning rate to be 0.001 and the weight attenuation coefficient to be 0.0001; setting the number of GPUs to be 1, the number of images processed by each GPU to be 2, and setting the receiving threshold value T of the damaged leaf0The other parameters are default values of Mask R-CNN;
(4.2) based on the known classification error LclsDetecting error LboxAnd a segmentation error LmaskAnd determining the optimized MaskR-CNN network loss function as follows: l isloss=Lcls+Lbox+Lmask;
(4.3) training the optimized Mask R-CNN network:
(4.3a) initializing the network parameters in (3a) and training the image set D5And training mask set D7Inputting the data into an optimized Mask R-CNN network;
(4.3b) extracting the characteristic map F of the training image through the training residual error network0;
(4.3c) mapping the feature pattern F0Inputting the information into a region recommendation generation network RPN to obtain the foreground F of the target region1And background F2;
(4.3d) reactingUsing ROI Align method to map the foreground F of the target area1Mapping to a feature map F0Generating a feature map F with a fixed size3:
First, the foreground F of the target area is calculated1The characteristic layer of the system is as follows: wherein ,w0 and h0Respectively representing the width and height of the target region, k0A value of 4;
second, in the target area foreground F1After finding out a corresponding characteristic layer k, obtaining a step length s corresponding to the characteristic layer;
then, the foreground F of the target area is calculated1Breadth mapping to feature mapAnd heightAnd obtaining a target area Z on the characteristic diagram according to the two parameters:
Z=w1*h1
next, the target region Z on the feature map is divided into n2Obtaining divided target areas Zi,
Zi=w2*h2,i=1,2,…n2
Then, each target region ZiDividing the image into four parts, obtaining pixel values of four points by taking the position of a central point of each part, and taking the maximum value of the four pixel values as each target area ZiTo obtain n in the target zone Z2Pixel values which form a characteristic map of n × n size;
(4.3e) feature map F3By full rollingStacking to obtain target classification result and target detection result, and calculating classification error LclsAnd the detection error Lbox:
Probability p corresponding to target classification result uuObtaining the classification error: l iscls=-logpu;
Let ti u={tx,ty,tw,thV 4 parameterized coordinates of the target detection result, vi={vx,vy,vw,vhCalculating the detection error L by the following formula as a target translation scaling parameterbox:
(4.3f) judging whether the classification result belongs to the diseased leaf:
if the leaf is not the diseased leaf, continuing to judge the target classification of the next leaf;
if the leaf is a diseased leaf, the confidence coefficient T of the classification result is determined1And a diseased leaf reception threshold T0Comparing;
when T is1>T0Then, in the feature diagram F3The confidence coefficient of the selected one is T1Characteristic diagram F of4;
When T is1<=T0If so, continuing to judge the target classification of the next blade;
(4.3g) feature map F selected in (4.3d)4And training mask set D7Inputting the binary mask into a mask branch for training to obtain a binary mask of a target area, namely a segmentation result of a disease leaf and a disease position of the disease leaf:
first, the feature map F is transformed by deconvolution3Amplifying to obtain binary mask regions M of all categoriesk;
Then, traverse all the categories of binary mask regions MkFor each class binary mask region MiApplying Sigmoid activation functionsAfter classification operation is carried out, a classification probability vector H is obtained, and the maximum probability y in the H is a binary mask of a target area corresponding to target classification;
(4.3h) calculating the segmentation error L of the binary mask of the target region obtained in (4.3e)mask;
Where y is the predicted probability of the binary mask of the target region,a real label of the binary mask of the target area;
(4.3i) calculating loss value L of networkloss=Lcls+Lbox+LmaskUpdating the network weight by back propagation of the loss value after each iteration;
(4.3g) judging whether the iteration times of all samples are more than the set 100 times:
if the iteration times of all samples are larger than 100, stopping network training to obtain a trained network model;
if the iteration number of all samples is less than or equal to 100, repeating the steps from (4.3c) to (4.3g) until the iteration number of all samples is more than 100.
And 5, obtaining the identification result of the plant leaf diseases.
(5.1) collecting the test image D obtained in (1.2)6Healthy leaves in (1) such as FIG. 5a and diseased leaves in (5 b) are input into the trained network model, feature vectors are extracted, and feature vectors A of each real target classification in n real leaf classes are calculatediCosine similarity cos with feature vector B of prediction target classificationi(θ), resulting in a probability vector P for the classification:
P={cos1(θ),cos2(θ),…,cosi(θ),…cosn(θ)},i=1,2…,n
wherein the cosine similarity cosiThe formula for (θ) is as follows:
wherein ,||AiThe | | represents the two-norm of the feature vector of each real target classification, and the | B | represents the two-norm of the feature vector of the prediction target classification;
(5.3) selecting the maximum probability value q in the probability vector P, and taking the category corresponding to q as a target classification result;
(5.2) judging whether the category corresponding to the q belongs to the diseased leaf:
if the class corresponding to q is a healthy leaf, directly outputting a target detection result and a target classification result;
if the category corresponding to q belongs to the damaged leaf, the q and the damaged leaf receiving threshold T are used0And (3) comparison:
if q > T0Outputting target detection, target classification and segmentation results of disease positions of the targets;
if q < ═ T0And directly outputting the target detection result and the target classification result.
The effect of the invention can be further illustrated by the following simulation experiment:
experimental conditions
The software platform for the experiment training is as follows: *** coliab; the hardware platform is as follows: tesla P4 GPU; the development environments are keras and Tensorflow;
the software platform for the experimental test is as follows: windows 10; the hardware platform is as follows: a CPU;
the test image is selected from (256, 3) size blade image shown in FIG. 5, and the test set is selected from the test set D obtained in step 23。
Second, the experimental contents
Experiment 1. A Mask R-CNN network model and the method of the invention are used for respectively carrying out comparison experiments on single test images.
Fig. 5 is a test image, in which fig. 5a is a healthy leaf and fig. 5b is a diseased leaf. The detection result of the Mask R-CNN network model is shown in FIG. 6, in which FIG. 6a is the detection result of healthy leaves, and FIG. 6b is the detection result of diseased leaves. The detection results of the method of the present invention are shown in fig. 7, wherein fig. 7a is the detection results of healthy leaves, and fig. 7b is the detection results of diseased leaves. From the detection results of the two methods, the method reduces redundant operation of dividing the diseased region of the blade compared with a MaskR-CNN network when detecting the healthy blade.
Experiment 2. comparative experiments were carried out with the method of the present invention and GoogleNet and VGG networks, respectively.
The method and the existing GoogleNet and VGG network models are used for carrying out the pair of test image sets D obtained in the step 161000 tests are carried out, the obtained identification accuracy rate results are shown in table 1,
TABLE 1 identification accuracy results
Method of producing a composite material | Average rate of accuracy | Time/s |
VGG | 0.8846 | 3.17 |
GoogleNet | 0.9040 | 3.32 |
Optimized Mask R-CNN | 0.9257 | 3.59 |
From table 1, it can be seen that the method of the present invention improves the recognition accuracy compared to the conventional GoogleNet and VGG network models.
The foregoing description is only a specific example of the present invention and is not intended to limit the invention, so that it will be apparent to those skilled in the art that modifications and variations in form and detail may be made without departing from the principles and structures of the invention, but such modifications and variations are within the scope of the invention as defined by the appended claims.
Claims (10)
1. A plant leaf disease identification method based on a mask convolution neural network is characterized by comprising the following steps:
(1) sequentially enhancing, expanding and semantically segmenting an original data set to obtain a training and testing image set and a mask set;
(2) optimizing a Mask R-CNN network: a disease characteristic screening module is added between the full convolution layer and the Mask branch after the Mask R-CNN network ROI Align;
(3) training the optimized Mask R-CNN network:
(3a) setting main network parameters:
selecting a main network from the ResNet50 and ResNet101 residual error networks;
setting the size of the epoch and the step number of each epoch training;
setting a receiving threshold T of a diseased leaf in a diseased feature screening module0The other parameters are default values of Mask R-CNN;
(3b) according to the known classification error LclsDetecting error LboxAnd a segmentation error LmaskAnd determining the optimized Mask R-CNN network loss function as follows: l isloss=Lcls+Lbox+Lmask;
(3c) Inputting the training image set and the training Mask set into an optimized Mask R-CNN network for training to obtain a trained model;
(4) and inputting the test image into the trained model for testing.
2. The method of claim 1, wherein (1) the original data set is sequentially enhanced, augmented and semantically segmented to obtain a trained and tested image set and a mask set, as follows:
(1a) the Plant disease leaf data set D is downloaded from the public project Plant Village-Dataset0Using adaptive contrast enhancement algorithm to pair D0Performing image enhancement to improve the blurred image in the database to obtain an image-enhanced data set D1;
(1b) Tagging tool for semantic segmentation on enhanced data set D1Respectively plotting the image targets to generate the mask of the targets, and obtaining a mask set D containing mask information and label information2;
(1c) Pair of enhanced data sets D with image transformations1Sum mask set D2Sequentially translating, rotating and overturning to increase the sample data volume and obtain a data set D after sample expansion3Sum mask set D4;
(1d) Scaling the augmented data set D by 8:23Sum mask set D4Divided into training image sets D5And testing the image set D4And a training mask set D7And test mask set D8。
3. The method of claim 2, wherein the image enhancement in (1a) is performed by using an adaptive contrast enhancement algorithm, and the following is performed:
(1a1) dividing the image into high-frequency parts hx(i, j) and a low frequency part mx(i, j), wherein (i, j) refers to pixel points of an image;
(1a2) multiplying the high frequency part in the image by the gain value G (I, j) of the part to obtain an amplified high frequency part I (I, j):
I(i,j)=G(i,j)hx(i,j)
(1a3) recombining the high frequency part and the low frequency part to obtain an enhanced image f (i, j):
f(i,j)=mx(i,j)+I(i,j)。
4. the method according to claim 1, wherein the optimized Mask R-CNN network obtained in (2) has the following structure:
residual error network ResNet101/50 or pyramid network FPN → area recommendation generation network RPN → full convolution layer → disease feature screening module → mask branch, namely full convolution layer and full connection layer;
the disease characteristic screening module is used for judging the confidence coefficient T of the disease blade output by the full convolution layer1With a given diseased leaf reception threshold T0Screening out the characteristic maps of the diseased leaves in batch processing, and inputting the screening result into a mask branch for processing.
5. The method according to claim 1, wherein the optimized Mask R-CNN network is trained in (3c) as follows:
(3c1) initializing the network parameters in (3a) and assembling the training image set D5And training mask set D7Inputting the data into the optimized MaskR-CNN network;
(3c2) extraction of D by training residual error network5To obtain a feature map F0;
(3c3) The characteristic map F0Inputting the information into a region recommendation generation network RPN to obtain the foreground F of the target region1And background F2;
(3c4) Foreground F of target area by using ROIAlign method1Mapping to a feature map F0Generating a feature map F with a fixed size3;
(3c5) Will feature chart F3Obtaining a target classification result and a target detection result through a full convolution layerAnd calculating a classification error LclsAnd the detection error Lbox;
(3c6) If the classification result is a diseased leaf, the confidence coefficient T of the classification result is determined1And a diseased leaf reception threshold T0By comparison, when T is1>T0Then, a disease characteristic screening module is utilized to screen a characteristic diagram F3The confidence coefficient of the selected one is T1Characteristic diagram F of4;
(3c7) The feature map F selected in (3c6)4And training mask set D7Inputting the binary mask into a mask branch for training to obtain a binary mask of a target area, namely a segmentation result of a disease leaf and a disease position of the disease leaf, and calculating a segmentation error Lmask;
(3c8) Calculating a loss value L of a networkloss=Lcls+Lbox+LmaskAnd updating the network weight by utilizing the back propagation of the loss value after each iteration, and stopping the network training when the training epoch is greater than the initialized epoch to obtain the trained network model.
6. The method of claim 5, wherein the ROI Align method is used to map the foreground F of the target region in (3c4)1Mapping to a feature map F0The corresponding positions of (a) are implemented as follows:
(3c4a) calculating a feature layer to which the target region belongs: wherein ,w0 and h0Respectively representing the width and height of the target region, k0A value of 4;
(3c4b) after finding the corresponding characteristic layer k in the target area, obtaining the step length s corresponding to the characteristic layer;
(3c4c) calculating the width w of the target region mapping to the feature map1And a height h1And obtaining a target area Z on the characteristic diagram:
Z=w1*h1
(3c4d) dividing the target zone Z on the feature map into n2Obtaining divided target areas Zi,
Zi=w2*h2,i=1,2,…n2
(3c4e) Each target zone ZiDividing the image into four parts, obtaining pixel values of four points by taking the position of a central point of each part, and taking the maximum value of the four pixel values as each target area ZiTo finally obtain n in the target zone Z2And (4) forming a characteristic map with the size of n × n by using the pixel values.
7. The method of claim 5, wherein the classification error L is calculated in (3c5)clsAnd the detection error LlocThe implementation is as follows:
(3c5a) probability p corresponding to the target classification result uuObtain a classification error Lcls:
Lcls=-logpu
(3c5b) setting ti u={tx,ty,tw,thV 4 parameterized coordinates of the target detection result, vi={vx,vy,vw,vhCalculating the detection error L by the following formula as a target translation scaling parameterbox:
8. the method of claim 5, wherein (3c7) the binary mask is obtained through a mask branch, and the following is implemented:
(3c7a) feature map F is transformed by deconvolution3Amplifying to obtain binary mask regions M of all categoriesk;
(3c7b) traversing all classes of binary mask regions MkFor each class binary mask region MiApplying Sigmoid activation functionsAfter classification operation is carried out, a classification probability vector H is obtained, and the maximum probability y in the H is a binary mask of a target area corresponding to target classification;
(3c7c) calculating the segmentation error L of the mask branchmask:
9. The method of claim 1, wherein the test image is input into the trained model for testing in (4) as follows:
(4a) collecting the test image D obtained in (1a)6Inputting the feature vectors into a trained network model, extracting the feature vectors, obtaining classified probability vectors P by calculating the similarity of the feature vectors, selecting the maximum probability value q in P, and taking the category corresponding to q as a target classification result;
(4b) judging whether the category corresponding to q belongs to a diseased leaf or not:
if the class corresponding to q is a healthy leaf, directly outputting a target detection result and a target classification result;
if the category corresponding to q belongs to the damaged leaf, the q and the damaged leaf receiving threshold T are used0And (3) comparison:
if q > T0Outputting target detection, target classification and segmentation results of disease positions;
if q < ═ T0And directly outputting the target detection result and the target classification result.
10. The method of claim 9, wherein the probability vector of the classification obtained by calculating the similarity of the feature vectors in (4a) is obtained by calculating the cosine similarity cos (θ) between the feature vector a of each real target classification and the feature vector B of the predicted target classification, and the formula is as follows:
wherein, | a | | represents the two-norm of the feature vector of the real target classification, and | B | | represents the two-norm of the feature vector of the prediction target classification.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010150980.9A CN111369540B (en) | 2020-03-06 | 2020-03-06 | Plant leaf disease identification method based on mask convolutional neural network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010150980.9A CN111369540B (en) | 2020-03-06 | 2020-03-06 | Plant leaf disease identification method based on mask convolutional neural network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN111369540A true CN111369540A (en) | 2020-07-03 |
CN111369540B CN111369540B (en) | 2023-06-02 |
Family
ID=71210311
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010150980.9A Active CN111369540B (en) | 2020-03-06 | 2020-03-06 | Plant leaf disease identification method based on mask convolutional neural network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111369540B (en) |
Cited By (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112115888A (en) * | 2020-09-22 | 2020-12-22 | 四川大学 | Plant disease diagnosis system based on disease spot correlation |
CN112560644A (en) * | 2020-12-11 | 2021-03-26 | 四川大学 | Crop disease and insect pest automatic identification method suitable for field |
CN112598031A (en) * | 2020-12-08 | 2021-04-02 | 北京农业信息技术研究中心 | Vegetable disease detection method and system |
CN112634147A (en) * | 2020-12-09 | 2021-04-09 | 上海健康医学院 | PET image noise reduction method, system, device and medium for self-supervision learning |
CN112699941A (en) * | 2020-12-31 | 2021-04-23 | 浙江科技学院 | Plant disease severity image classification method and device, computer equipment and storage medium |
CN112884022A (en) * | 2021-01-29 | 2021-06-01 | 浙江师范大学 | Unsupervised depth characterization learning method and system based on image translation |
CN113052799A (en) * | 2021-03-09 | 2021-06-29 | 重庆大学 | Osteosarcoma and osteochondroma prediction method based on Mask RCNN network |
CN113112498A (en) * | 2021-05-06 | 2021-07-13 | 东北农业大学 | Grape leaf scab identification method based on fine-grained countermeasure generation network |
CN113191334A (en) * | 2021-05-31 | 2021-07-30 | 广西师范大学 | Plant canopy dense leaf counting method based on improved CenterNet |
CN113239788A (en) * | 2021-05-11 | 2021-08-10 | 嘉兴学院 | Mask R-CNN-based wireless communication modulation mode identification method |
CN113762190A (en) * | 2021-09-15 | 2021-12-07 | 中科微至智能制造科技江苏股份有限公司 | Neural network-based parcel stacking detection method and device |
CN114239756A (en) * | 2022-02-25 | 2022-03-25 | 科大天工智能装备技术(天津)有限公司 | Insect pest detection method and system |
CN114742204A (en) * | 2022-04-08 | 2022-07-12 | 黑龙江惠达科技发展有限公司 | Method and device for detecting straw coverage rate |
CN114943988A (en) * | 2022-06-16 | 2022-08-26 | 浙大城市学院 | Planar target monitoring method based on instance segmentation and deep convolutional neural network |
CN117011718A (en) * | 2023-10-08 | 2023-11-07 | 之江实验室 | Plant leaf fine granularity identification method and system based on multiple loss fusion |
CN117372790A (en) * | 2023-12-08 | 2024-01-09 | 浙江托普云农科技股份有限公司 | Plant leaf shape classification method, system and device |
CN117882546A (en) * | 2024-03-13 | 2024-04-16 | 山西诚鼎伟业科技有限责任公司 | Intelligent planting method for agricultural operation robot |
CN118072177A (en) * | 2024-04-12 | 2024-05-24 | 国网江西省电力有限公司电力科学研究院 | Line channel inflammable tree species identification method based on laser point cloud and image fusion |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190065817A1 (en) * | 2017-08-29 | 2019-02-28 | Konica Minolta Laboratory U.S.A., Inc. | Method and system for detection and classification of cells using convolutional neural networks |
CN110163207A (en) * | 2019-05-20 | 2019-08-23 | 福建船政交通职业学院 | One kind is based on Mask-RCNN ship target localization method and storage equipment |
CN110717903A (en) * | 2019-09-30 | 2020-01-21 | 天津大学 | Method for detecting crop diseases by using computer vision technology |
-
2020
- 2020-03-06 CN CN202010150980.9A patent/CN111369540B/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20190065817A1 (en) * | 2017-08-29 | 2019-02-28 | Konica Minolta Laboratory U.S.A., Inc. | Method and system for detection and classification of cells using convolutional neural networks |
CN110163207A (en) * | 2019-05-20 | 2019-08-23 | 福建船政交通职业学院 | One kind is based on Mask-RCNN ship target localization method and storage equipment |
CN110717903A (en) * | 2019-09-30 | 2020-01-21 | 天津大学 | Method for detecting crop diseases by using computer vision technology |
Non-Patent Citations (1)
Title |
---|
王艳玲等: "基于迁移学习的番茄叶片病害图像分类", 《中国农业大学学报》 * |
Cited By (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112115888B (en) * | 2020-09-22 | 2022-06-03 | 四川大学 | Plant disease diagnosis system based on disease spot correlation |
CN112115888A (en) * | 2020-09-22 | 2020-12-22 | 四川大学 | Plant disease diagnosis system based on disease spot correlation |
CN112598031A (en) * | 2020-12-08 | 2021-04-02 | 北京农业信息技术研究中心 | Vegetable disease detection method and system |
CN112634147A (en) * | 2020-12-09 | 2021-04-09 | 上海健康医学院 | PET image noise reduction method, system, device and medium for self-supervision learning |
CN112634147B (en) * | 2020-12-09 | 2024-03-29 | 上海健康医学院 | PET image noise reduction method, system, device and medium for self-supervision learning |
CN112560644A (en) * | 2020-12-11 | 2021-03-26 | 四川大学 | Crop disease and insect pest automatic identification method suitable for field |
CN112560644B (en) * | 2020-12-11 | 2021-09-28 | 四川大学 | Crop disease and insect pest automatic identification method suitable for field |
CN112699941A (en) * | 2020-12-31 | 2021-04-23 | 浙江科技学院 | Plant disease severity image classification method and device, computer equipment and storage medium |
CN112699941B (en) * | 2020-12-31 | 2023-02-14 | 浙江科技学院 | Plant disease severity image classification method, device, equipment and storage medium |
CN112884022A (en) * | 2021-01-29 | 2021-06-01 | 浙江师范大学 | Unsupervised depth characterization learning method and system based on image translation |
CN113052799A (en) * | 2021-03-09 | 2021-06-29 | 重庆大学 | Osteosarcoma and osteochondroma prediction method based on Mask RCNN network |
CN113112498A (en) * | 2021-05-06 | 2021-07-13 | 东北农业大学 | Grape leaf scab identification method based on fine-grained countermeasure generation network |
CN113112498B (en) * | 2021-05-06 | 2024-01-19 | 东北农业大学 | Grape leaf spot identification method based on fine-grained countermeasure generation network |
CN113239788A (en) * | 2021-05-11 | 2021-08-10 | 嘉兴学院 | Mask R-CNN-based wireless communication modulation mode identification method |
CN113191334B (en) * | 2021-05-31 | 2022-07-01 | 广西师范大学 | Plant canopy dense leaf counting method based on improved CenterNet |
CN113191334A (en) * | 2021-05-31 | 2021-07-30 | 广西师范大学 | Plant canopy dense leaf counting method based on improved CenterNet |
CN113762190A (en) * | 2021-09-15 | 2021-12-07 | 中科微至智能制造科技江苏股份有限公司 | Neural network-based parcel stacking detection method and device |
CN113762190B (en) * | 2021-09-15 | 2024-03-29 | 中科微至科技股份有限公司 | Method and device for detecting package stacking based on neural network |
CN114239756B (en) * | 2022-02-25 | 2022-05-17 | 科大天工智能装备技术(天津)有限公司 | Insect pest detection method and system |
CN114239756A (en) * | 2022-02-25 | 2022-03-25 | 科大天工智能装备技术(天津)有限公司 | Insect pest detection method and system |
CN114742204A (en) * | 2022-04-08 | 2022-07-12 | 黑龙江惠达科技发展有限公司 | Method and device for detecting straw coverage rate |
CN114943988A (en) * | 2022-06-16 | 2022-08-26 | 浙大城市学院 | Planar target monitoring method based on instance segmentation and deep convolutional neural network |
CN114943988B (en) * | 2022-06-16 | 2024-04-02 | 浙大城市学院 | Planar target monitoring method based on instance segmentation and deep convolution neural network |
CN117011718B (en) * | 2023-10-08 | 2024-02-02 | 之江实验室 | Plant leaf fine granularity identification method and system based on multiple loss fusion |
CN117011718A (en) * | 2023-10-08 | 2023-11-07 | 之江实验室 | Plant leaf fine granularity identification method and system based on multiple loss fusion |
CN117372790A (en) * | 2023-12-08 | 2024-01-09 | 浙江托普云农科技股份有限公司 | Plant leaf shape classification method, system and device |
CN117372790B (en) * | 2023-12-08 | 2024-03-08 | 浙江托普云农科技股份有限公司 | Plant leaf shape classification method, system and device |
CN117882546A (en) * | 2024-03-13 | 2024-04-16 | 山西诚鼎伟业科技有限责任公司 | Intelligent planting method for agricultural operation robot |
CN117882546B (en) * | 2024-03-13 | 2024-05-24 | 山西诚鼎伟业科技有限责任公司 | Intelligent planting method for agricultural operation robot |
CN118072177A (en) * | 2024-04-12 | 2024-05-24 | 国网江西省电力有限公司电力科学研究院 | Line channel inflammable tree species identification method based on laser point cloud and image fusion |
Also Published As
Publication number | Publication date |
---|---|
CN111369540B (en) | 2023-06-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111369540B (en) | Plant leaf disease identification method based on mask convolutional neural network | |
Dias et al. | Multispecies fruit flower detection using a refined semantic segmentation network | |
WO2022160771A1 (en) | Method for classifying hyperspectral images on basis of adaptive multi-scale feature extraction model | |
CN109344883A (en) | Fruit tree diseases and pests recognition methods under a kind of complex background based on empty convolution | |
CN110222215B (en) | Crop pest detection method based on F-SSD-IV3 | |
CN111860330A (en) | Apple leaf disease identification method based on multi-feature fusion and convolutional neural network | |
CN110176005B (en) | Remote sensing image segmentation method based on normalized index and multi-scale model | |
CN110853070A (en) | Underwater sea cucumber image segmentation method based on significance and Grabcut | |
US11037299B2 (en) | Region merging image segmentation algorithm based on boundary extraction | |
CN112749675A (en) | Potato disease identification method based on convolutional neural network | |
Chen et al. | Segmentation of field grape bunches via an improved pyramid scene parsing network | |
Zhang et al. | Robust image segmentation method for cotton leaf under natural conditions based on immune algorithm and PCNN algorithm | |
CN113569772A (en) | Remote sensing image farmland instance mask extraction method, system, equipment and storage medium | |
CN113627481A (en) | Multi-model combined unmanned aerial vehicle garbage classification method for smart gardens | |
CN111832480B (en) | Remote sensing identification method for rape planting area based on spectral characteristics | |
CN113077452A (en) | Apple tree pest and disease detection method based on DNN network and spot detection algorithm | |
Wang et al. | MRUNet: A two-stage segmentation model for small insect targets in complex environments | |
CN114596502B (en) | Arable land information extraction method based on edge-enhanced high-resolution remote sensing image | |
CN113723833B (en) | Method, system, terminal equipment and storage medium for evaluating quality of forestation actual results | |
CN114663791A (en) | Branch recognition method for pruning robot in unstructured environment | |
Rony et al. | BottleNet18: Deep Learning-Based Bottle Gourd Leaf Disease Classification | |
CN115601634A (en) | Image blade identification method and device based on hierarchical attention mechanism | |
Fan et al. | Deep neural de-raining model based on dynamic fusion of multiple vision tasks | |
CN111046861A (en) | Method for identifying whether sample equipment exists in infrared image, method for constructing power equipment identification model and application | |
Song et al. | Crop Classification Using Multitemporal Landsat 8 Images |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |