CN111369540A - Plant leaf disease identification method based on mask convolutional neural network - Google Patents

Plant leaf disease identification method based on mask convolutional neural network Download PDF

Info

Publication number
CN111369540A
CN111369540A CN202010150980.9A CN202010150980A CN111369540A CN 111369540 A CN111369540 A CN 111369540A CN 202010150980 A CN202010150980 A CN 202010150980A CN 111369540 A CN111369540 A CN 111369540A
Authority
CN
China
Prior art keywords
mask
target
network
image
training
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010150980.9A
Other languages
Chinese (zh)
Other versions
CN111369540B (en
Inventor
王勇
刘雪月
胥克翔
靳伟昭
杨琦
朱文涛
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN202010150980.9A priority Critical patent/CN111369540B/en
Publication of CN111369540A publication Critical patent/CN111369540A/en
Application granted granted Critical
Publication of CN111369540B publication Critical patent/CN111369540B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/73Deblurring; Sharpening
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30181Earth observation
    • G06T2207/30188Vegetation; Agriculture
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A40/00Adaptation technologies in agriculture, forestry, livestock or agroalimentary production
    • Y02A40/10Adaptation technologies in agriculture, forestry, livestock or agroalimentary production in agriculture

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Data Mining & Analysis (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Quality & Reliability (AREA)
  • Multimedia (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a plant leaf disease identification method based on a mask convolution neural network, which mainly solves the problem of low accuracy rate of identifying plant leaf diseases in the prior art. The scheme is as follows: enhancing and expanding the original data set to obtain a training set and a test set; performing semantic segmentation on the training set and the test set to obtain a corresponding mask set; adding a disease characteristic screening module between the full convolution layer and the mask branch of the model, inputting the training set and the mask set into a network for training, and obtaining the results of target classification and target detection; taking a characteristic graph belonging to disease leaves in a target classification result as the input of a mask branch, and obtaining a trained model after multiple iterations; and inputting the test set into the model, performing target classification and target detection on the leaves, and segmenting the leaves belonging to the disease category. The method improves the accuracy of leaf disease identification on the basis of the traditional method, and can be used for identifying and segmenting plant disease leaves in agricultural planting.

Description

Plant leaf disease identification method based on mask convolutional neural network
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a plant leaf disease identification method which can be used for segmenting and identifying plant disease leaves in agricultural planting.
Background
In modern wisdom agriculture, crop diseases are a great threat to grain safety, and plant diseases cause serious damage to crops by significantly reducing yield. Early blight is a typical disease that can severely reduce yield. Similarly, in humid climates, late blight is another very devastating disease that affects the leaves, stems and fruits of plants. Protection of plants from disease is critical to ensure the quality and quantity of crops. Crop protection should begin with the early detection of disease so that appropriate treatments can be selected at the right time to prevent disease transmission. The method for identifying the disease types of the greenhouse plant diseases mainly comprises various diseases such as bacterial spots, early blight, late blight, leaf mold, black spot and the like. However, in practice, the number of diseases is large, and the diseases are similar in performance on leaves, so that it is difficult to accurately determine the types of the diseases.
At present, the disease research on plant leaves mainly comprises detecting and classifying the plant leaves by using an image processing or deep learning method. In the Plant disease control, a foreign Plant Village group provides a method for detecting and classifying diseases through Plant leaves by utilizing deep learning in 2016, mainly classifies specific partial Plant diseases through a smart phone under a simple background, and mainly identifies colors, gray scales and segmented pictures through the deep learning method under data sets with different proportions and different networks, but the method can only classify the Plant disease leaves and cannot segment the Plant disease leaves and the positions of the diseases.
The same year Mads Dyrmann1 et al proposed the use of convolutional neural networks to classify plant species in color images. The network is built from scratch, trained and tested. These images are from six different local datasets, with data acquisition at different growth stages, in terms of illumination, resolution and soil type, but with low accuracy and without detection and classification of plant diseases.
Liuna et al in 2018 use an image processing technology and an artificial neural network technology to realize disease detection and disease degree classification of cucumber leaves, and mainly perform experimental research on cucumber downy mildew, powdery mildew and virus diseases with high morbidity and serious harm, but because the disease types of cucumbers are identified less, the number of training samples is less, and the identified disease types are less, the test accuracy is lower and overfitting is easily generated.
The framework of Mask convolutional neural network Mask R-CNN is provided by Heterokamm in 2017, and is characterized in that semantic segmentation is carried out on two branches of FasterR-CNN, namely, on classification and coordinate regression, one branch is added, image features are extracted mainly through a residual error network ResNet101/50 or a pyramid network FPN serving as a main network, the foreground and the background of a target region are obtained through a region recommendation network RPN, classification and example segmentation results are obtained through full convolutional layers on the obtained target region and the image features, and then the semantic segmentation results are obtained through convolutional network identification. The network is mainly used for target detection of a COCO data set, and is not used for the field of plant disease identification at present.
In summary, the current research on plant diseases mainly includes classification and identification of single plant disease types similar to cucumber, wherein the included samples include fewer disease types and samples, and the identification accuracy is low. The existing method for classifying the plant disease leaves through deep learning cannot separate the plant disease leaves and the positions of diseases, and the recognition rate is low.
Disclosure of Invention
The invention aims to provide a plant leaf disease identification method based on mask convolution neural network MaskR-CNN aiming at the defects in the prior art, so as to segment plant disease leaves and the positions of diseases, and improve the identification accuracy and efficiency.
In order to achieve the purpose, the technical scheme of the invention comprises the following steps:
(1) sequentially enhancing, expanding and semantically segmenting an original data set to obtain a training and testing image set and a mask set;
(2) optimizing a Mask R-CNN network: a disease characteristic screening module is added between the full convolution layer and the Mask branch after the Mask R-CNN network ROI Align;
(3) training the optimized Mask R-CNN network:
(3a) setting main network parameters:
selecting a main network from the ResNet50 and ResNet101 residual error networks;
setting the size of the epoch and the step number of each epoch training;
setting a receiving threshold T of a diseased leaf in a diseased feature screening module0The other parameters are default values of Mask R-CNN;
(3b) according to the known classification error LclsDetecting error LboxAnd a segmentation error LmaskAnd determining the optimized MaskR-CNN network loss function as follows: l isloss=Lcls+Lbox+Lmask
(3c) Inputting the training image set and the training Mask set into an optimized Mask R-CNN network for training to obtain a trained model;
(4) and inputting the test image into the trained model for testing.
Compared with the prior art, the invention has the following advantages:
firstly, compared with the GoogleNet and VGG methods, the method disclosed by the invention has the advantage that the identification accuracy of the damaged leaves is improved when the plant leaf diseases are identified.
Secondly, the Mask image of the target area is generated through the Mask branch of the Mask R-CNN network model, so that the plant disease leaves and the disease positions thereof can be accurately extracted.
Thirdly, because the disease characteristic screening module is added in the network structure of the Mask R-CNN, the invention only aims at unhealthy leaves to carry out the training and testing of Mask branches, reduces the burden of the Mask branches and improves the identification efficiency of the network on the premise of ensuring that the accuracy of the Mask R-CNN network is not changed.
Fourthly, in the invention, because the samples in the data set are increased by adopting image transformation and the data set is enhanced by adopting an adaptive contrast enhancement algorithm, the blurred image in the data set is improved.
Drawings
FIG. 1 is a general flow diagram of an implementation of the present invention;
FIG. 2 is a schematic diagram of the overall structure of an optimized Mask R-CNN network in the present invention;
FIG. 3 is a block diagram of a training sub-process for an optimized Mask R-CNN network in the present invention;
FIG. 4 is a training image and a training binary mask image acquired in the present invention;
FIG. 5 is a test image of healthy and diseased leaves taken in the present invention;
FIG. 6 is a result image of healthy leaves identified by simulation of the present invention;
FIG. 7 is a result image of diseased leaves identified by simulation in accordance with the present invention.
Detailed Description
The following detailed description of the embodiments and effects of the present invention will be made with reference to the accompanying drawings:
the application environment of the embodiment is an agricultural planting scene, and aims to detect and identify vegetation with diseases in agricultural planting and provide disease information for planting personnel more accurately.
Referring to fig. 1, the implementation steps of this example are as follows:
step 1, image enhanced data set D1
(1.1) plant disease leaf Dataset D was downloaded from public project PlantVillage-Dataset0Using adaptive contrast enhancement algorithm to pair D0Performing image enhancement to improve the blurred image in the database to obtain an image-enhanced database D1
(1.1a) acquiring the low frequency part m of the image x (i, j)x(i, j) and a high frequency part hx(i, j), obtaining the low frequency part of the image by means filtering:
Figure BDA0002402432800000031
hx(i,j)=x(i,j)-mx(i,j)
wherein, (2n +1)2(ii) represents the window size with (i, j) as the image center point coordinate;
(1.1b) multiplying the high frequency part in the image by the gain value G (I, j) of the part to obtain an amplified high frequency part I (I, j):
I(i,j)=G(i,j)hx(i,j)
wherein G (I, j) may take a constant C greater than 1 to obtain an amplified high-frequency portion Ic(i, j) is:
Ic(i,j)=Chx(i,j)
in this example the gain value G (i, j) is taken to be equal to the local mean square error σx(i, j) is inversely proportional to the change value
Figure BDA0002402432800000041
Wherein D is a constant, and the local mean square error of the image is:
Figure BDA0002402432800000042
obtaining an amplified high-frequency part Iσ(i, j) is:
Figure BDA0002402432800000043
(1.1c) recombining the high frequency part and the low frequency part to obtain an enhanced image f (i, j):
f(i,j)=mx(i,j)+Iσ(i,j);
step 2, obtaining a training image set D3And testing the image set D4And a training mask set D5And test mask set D6
(2.1) enhancing the image data set D by using a semantic segmentation labeling tool labelme1Respectively plotting the image targets to generate the mask of the targets, and obtaining a mask set D containing mask information and label information2The resulting training binary mask image is shown in fig. 4 b.
(2.2) Using image transformation on the enhanced dataset D1Sum mask set D2Sequentially translating, rotating and overturning to increase the sample data volume and obtain a data set D after sample expansion3Sum mask set D4
(2.2) expanding the data set D3Sum mask set D4Dividing the training image into training image sets D according to the proportion of 8:25And testing the image set D6And a training mask set D7And test mask set D8The obtained training set image is shown in fig. 4a, and the test set image is shown in fig. 5a and 5 b;
and step 3, continuously optimizing the Mask R-CNN network structure.
The existing Mask R-CNN network comprises a main network, a regional recommendation generation network RPN, a full convolution layer and a Mask branch, namely the full convolution layer and a full connection layer, in the embodiment, a disease characteristic screening module is added between the full convolution layer and the Mask branch to obtain an optimized Mask R-CNN network structure, which is shown in figure 2.
Referring to fig. 2, the Mask R-CNN network structure after optimization in this example is: the main network → the area recommendation generation network RPN → the full convolution layer → the disease feature screening module → the full convolution layer → the full connection layer.
The disease characteristic screening moduleThe block is to judge the confidence coefficient T of the fault blade output by the full convolution layer1Receiving threshold T of damaged leaf given by network initialization parameter0The characteristic maps of the diseased leaves in batch processing are screened out, and screening results are input to mask branches, namely, full convolution layers and full connection for processing.
Step 4, training the optimized Mask R-CNN network to obtain a trained network model:
referring to fig. 3, the specific implementation of this step is as follows:
(4.1) setting main network parameters:
selecting a main network from the ResNet50 and ResNet101 residual error networks, wherein ResNet101 is selected as the main network in the example;
setting the number of labels to be 11 according to the category of the images in the database, wherein the number of labels comprises 1 background label and 10 image labels;
setting the iteration times epoch of all samples to be 100, the iteration times of each epoch to be 100, the learning rate to be 0.001 and the weight attenuation coefficient to be 0.0001; setting the number of GPUs to be 1, the number of images processed by each GPU to be 2, and setting the receiving threshold value T of the damaged leaf0The other parameters are default values of Mask R-CNN;
(4.2) based on the known classification error LclsDetecting error LboxAnd a segmentation error LmaskAnd determining the optimized MaskR-CNN network loss function as follows: l isloss=Lcls+Lbox+Lmask
(4.3) training the optimized Mask R-CNN network:
(4.3a) initializing the network parameters in (3a) and training the image set D5And training mask set D7Inputting the data into an optimized Mask R-CNN network;
(4.3b) extracting the characteristic map F of the training image through the training residual error network0
(4.3c) mapping the feature pattern F0Inputting the information into a region recommendation generation network RPN to obtain the foreground F of the target region1And background F2
(4.3d) reactingUsing ROI Align method to map the foreground F of the target area1Mapping to a feature map F0Generating a feature map F with a fixed size3
First, the foreground F of the target area is calculated1The characteristic layer of the system is as follows:
Figure BDA0002402432800000061
wherein ,w0 and h0Respectively representing the width and height of the target region, k0A value of 4;
second, in the target area foreground F1After finding out a corresponding characteristic layer k, obtaining a step length s corresponding to the characteristic layer;
then, the foreground F of the target area is calculated1Breadth mapping to feature map
Figure BDA0002402432800000062
And height
Figure BDA0002402432800000063
And obtaining a target area Z on the characteristic diagram according to the two parameters:
Z=w1*h1
next, the target region Z on the feature map is divided into n2Obtaining divided target areas Zi
Zi=w2*h2,i=1,2,…n2
wherein ,w2 and h2Represents ZiWidth and height, respectively
Figure BDA0002402432800000064
Then, each target region ZiDividing the image into four parts, obtaining pixel values of four points by taking the position of a central point of each part, and taking the maximum value of the four pixel values as each target area ZiTo obtain n in the target zone Z2Pixel values which form a characteristic map of n × n size;
(4.3e) feature map F3By full rollingStacking to obtain target classification result and target detection result, and calculating classification error LclsAnd the detection error Lbox
Probability p corresponding to target classification result uuObtaining the classification error: l iscls=-logpu
Let ti u={tx,ty,tw,thV 4 parameterized coordinates of the target detection result, vi={vx,vy,vw,vhCalculating the detection error L by the following formula as a target translation scaling parameterbox
Figure BDA0002402432800000065
Wherein smoothL1Is a smoothed norm loss function expressed as:
Figure BDA0002402432800000071
(4.3f) judging whether the classification result belongs to the diseased leaf:
if the leaf is not the diseased leaf, continuing to judge the target classification of the next leaf;
if the leaf is a diseased leaf, the confidence coefficient T of the classification result is determined1And a diseased leaf reception threshold T0Comparing;
when T is1>T0Then, in the feature diagram F3The confidence coefficient of the selected one is T1Characteristic diagram F of4
When T is1<=T0If so, continuing to judge the target classification of the next blade;
(4.3g) feature map F selected in (4.3d)4And training mask set D7Inputting the binary mask into a mask branch for training to obtain a binary mask of a target area, namely a segmentation result of a disease leaf and a disease position of the disease leaf:
first, the feature map F is transformed by deconvolution3Amplifying to obtain binary mask regions M of all categoriesk
Then, traverse all the categories of binary mask regions MkFor each class binary mask region MiApplying Sigmoid activation functions
Figure BDA0002402432800000072
After classification operation is carried out, a classification probability vector H is obtained, and the maximum probability y in the H is a binary mask of a target area corresponding to target classification;
(4.3h) calculating the segmentation error L of the binary mask of the target region obtained in (4.3e)mask
Figure BDA0002402432800000073
Where y is the predicted probability of the binary mask of the target region,
Figure BDA0002402432800000074
a real label of the binary mask of the target area;
(4.3i) calculating loss value L of networkloss=Lcls+Lbox+LmaskUpdating the network weight by back propagation of the loss value after each iteration;
(4.3g) judging whether the iteration times of all samples are more than the set 100 times:
if the iteration times of all samples are larger than 100, stopping network training to obtain a trained network model;
if the iteration number of all samples is less than or equal to 100, repeating the steps from (4.3c) to (4.3g) until the iteration number of all samples is more than 100.
And 5, obtaining the identification result of the plant leaf diseases.
(5.1) collecting the test image D obtained in (1.2)6Healthy leaves in (1) such as FIG. 5a and diseased leaves in (5 b) are input into the trained network model, feature vectors are extracted, and feature vectors A of each real target classification in n real leaf classes are calculatediCosine similarity cos with feature vector B of prediction target classificationi(θ), resulting in a probability vector P for the classification:
P={cos1(θ),cos2(θ),…,cosi(θ),…cosn(θ)},i=1,2…,n
wherein the cosine similarity cosiThe formula for (θ) is as follows:
Figure BDA0002402432800000081
wherein ,||AiThe | | represents the two-norm of the feature vector of each real target classification, and the | B | represents the two-norm of the feature vector of the prediction target classification;
(5.3) selecting the maximum probability value q in the probability vector P, and taking the category corresponding to q as a target classification result;
(5.2) judging whether the category corresponding to the q belongs to the diseased leaf:
if the class corresponding to q is a healthy leaf, directly outputting a target detection result and a target classification result;
if the category corresponding to q belongs to the damaged leaf, the q and the damaged leaf receiving threshold T are used0And (3) comparison:
if q > T0Outputting target detection, target classification and segmentation results of disease positions of the targets;
if q < ═ T0And directly outputting the target detection result and the target classification result.
The effect of the invention can be further illustrated by the following simulation experiment:
experimental conditions
The software platform for the experiment training is as follows: *** coliab; the hardware platform is as follows: tesla P4 GPU; the development environments are keras and Tensorflow;
the software platform for the experimental test is as follows: windows 10; the hardware platform is as follows: a CPU;
the test image is selected from (256, 3) size blade image shown in FIG. 5, and the test set is selected from the test set D obtained in step 23
Second, the experimental contents
Experiment 1. A Mask R-CNN network model and the method of the invention are used for respectively carrying out comparison experiments on single test images.
Fig. 5 is a test image, in which fig. 5a is a healthy leaf and fig. 5b is a diseased leaf. The detection result of the Mask R-CNN network model is shown in FIG. 6, in which FIG. 6a is the detection result of healthy leaves, and FIG. 6b is the detection result of diseased leaves. The detection results of the method of the present invention are shown in fig. 7, wherein fig. 7a is the detection results of healthy leaves, and fig. 7b is the detection results of diseased leaves. From the detection results of the two methods, the method reduces redundant operation of dividing the diseased region of the blade compared with a MaskR-CNN network when detecting the healthy blade.
Experiment 2. comparative experiments were carried out with the method of the present invention and GoogleNet and VGG networks, respectively.
The method and the existing GoogleNet and VGG network models are used for carrying out the pair of test image sets D obtained in the step 161000 tests are carried out, the obtained identification accuracy rate results are shown in table 1,
TABLE 1 identification accuracy results
Method of producing a composite material Average rate of accuracy Time/s
VGG 0.8846 3.17
GoogleNet 0.9040 3.32
Optimized Mask R-CNN 0.9257 3.59
From table 1, it can be seen that the method of the present invention improves the recognition accuracy compared to the conventional GoogleNet and VGG network models.
The foregoing description is only a specific example of the present invention and is not intended to limit the invention, so that it will be apparent to those skilled in the art that modifications and variations in form and detail may be made without departing from the principles and structures of the invention, but such modifications and variations are within the scope of the invention as defined by the appended claims.

Claims (10)

1. A plant leaf disease identification method based on a mask convolution neural network is characterized by comprising the following steps:
(1) sequentially enhancing, expanding and semantically segmenting an original data set to obtain a training and testing image set and a mask set;
(2) optimizing a Mask R-CNN network: a disease characteristic screening module is added between the full convolution layer and the Mask branch after the Mask R-CNN network ROI Align;
(3) training the optimized Mask R-CNN network:
(3a) setting main network parameters:
selecting a main network from the ResNet50 and ResNet101 residual error networks;
setting the size of the epoch and the step number of each epoch training;
setting a receiving threshold T of a diseased leaf in a diseased feature screening module0The other parameters are default values of Mask R-CNN;
(3b) according to the known classification error LclsDetecting error LboxAnd a segmentation error LmaskAnd determining the optimized Mask R-CNN network loss function as follows: l isloss=Lcls+Lbox+Lmask
(3c) Inputting the training image set and the training Mask set into an optimized Mask R-CNN network for training to obtain a trained model;
(4) and inputting the test image into the trained model for testing.
2. The method of claim 1, wherein (1) the original data set is sequentially enhanced, augmented and semantically segmented to obtain a trained and tested image set and a mask set, as follows:
(1a) the Plant disease leaf data set D is downloaded from the public project Plant Village-Dataset0Using adaptive contrast enhancement algorithm to pair D0Performing image enhancement to improve the blurred image in the database to obtain an image-enhanced data set D1
(1b) Tagging tool for semantic segmentation on enhanced data set D1Respectively plotting the image targets to generate the mask of the targets, and obtaining a mask set D containing mask information and label information2
(1c) Pair of enhanced data sets D with image transformations1Sum mask set D2Sequentially translating, rotating and overturning to increase the sample data volume and obtain a data set D after sample expansion3Sum mask set D4
(1d) Scaling the augmented data set D by 8:23Sum mask set D4Divided into training image sets D5And testing the image set D4And a training mask set D7And test mask set D8
3. The method of claim 2, wherein the image enhancement in (1a) is performed by using an adaptive contrast enhancement algorithm, and the following is performed:
(1a1) dividing the image into high-frequency parts hx(i, j) and a low frequency part mx(i, j), wherein (i, j) refers to pixel points of an image;
(1a2) multiplying the high frequency part in the image by the gain value G (I, j) of the part to obtain an amplified high frequency part I (I, j):
I(i,j)=G(i,j)hx(i,j)
(1a3) recombining the high frequency part and the low frequency part to obtain an enhanced image f (i, j):
f(i,j)=mx(i,j)+I(i,j)。
4. the method according to claim 1, wherein the optimized Mask R-CNN network obtained in (2) has the following structure:
residual error network ResNet101/50 or pyramid network FPN → area recommendation generation network RPN → full convolution layer → disease feature screening module → mask branch, namely full convolution layer and full connection layer;
the disease characteristic screening module is used for judging the confidence coefficient T of the disease blade output by the full convolution layer1With a given diseased leaf reception threshold T0Screening out the characteristic maps of the diseased leaves in batch processing, and inputting the screening result into a mask branch for processing.
5. The method according to claim 1, wherein the optimized Mask R-CNN network is trained in (3c) as follows:
(3c1) initializing the network parameters in (3a) and assembling the training image set D5And training mask set D7Inputting the data into the optimized MaskR-CNN network;
(3c2) extraction of D by training residual error network5To obtain a feature map F0
(3c3) The characteristic map F0Inputting the information into a region recommendation generation network RPN to obtain the foreground F of the target region1And background F2
(3c4) Foreground F of target area by using ROIAlign method1Mapping to a feature map F0Generating a feature map F with a fixed size3
(3c5) Will feature chart F3Obtaining a target classification result and a target detection result through a full convolution layerAnd calculating a classification error LclsAnd the detection error Lbox
(3c6) If the classification result is a diseased leaf, the confidence coefficient T of the classification result is determined1And a diseased leaf reception threshold T0By comparison, when T is1>T0Then, a disease characteristic screening module is utilized to screen a characteristic diagram F3The confidence coefficient of the selected one is T1Characteristic diagram F of4
(3c7) The feature map F selected in (3c6)4And training mask set D7Inputting the binary mask into a mask branch for training to obtain a binary mask of a target area, namely a segmentation result of a disease leaf and a disease position of the disease leaf, and calculating a segmentation error Lmask
(3c8) Calculating a loss value L of a networkloss=Lcls+Lbox+LmaskAnd updating the network weight by utilizing the back propagation of the loss value after each iteration, and stopping the network training when the training epoch is greater than the initialized epoch to obtain the trained network model.
6. The method of claim 5, wherein the ROI Align method is used to map the foreground F of the target region in (3c4)1Mapping to a feature map F0The corresponding positions of (a) are implemented as follows:
(3c4a) calculating a feature layer to which the target region belongs:
Figure FDA0002402432790000031
wherein ,w0 and h0Respectively representing the width and height of the target region, k0A value of 4;
(3c4b) after finding the corresponding characteristic layer k in the target area, obtaining the step length s corresponding to the characteristic layer;
(3c4c) calculating the width w of the target region mapping to the feature map1And a height h1And obtaining a target area Z on the characteristic diagram:
Z=w1*h1
wherein ,
Figure FDA0002402432790000032
(3c4d) dividing the target zone Z on the feature map into n2Obtaining divided target areas Zi
Zi=w2*h2,i=1,2,…n2
wherein ,w2 and h2Represents ZiWidth and height, respectively
Figure FDA0002402432790000041
(3c4e) Each target zone ZiDividing the image into four parts, obtaining pixel values of four points by taking the position of a central point of each part, and taking the maximum value of the four pixel values as each target area ZiTo finally obtain n in the target zone Z2And (4) forming a characteristic map with the size of n × n by using the pixel values.
7. The method of claim 5, wherein the classification error L is calculated in (3c5)clsAnd the detection error LlocThe implementation is as follows:
(3c5a) probability p corresponding to the target classification result uuObtain a classification error Lcls
Lcls=-logpu
(3c5b) setting ti u={tx,ty,tw,thV 4 parameterized coordinates of the target detection result, vi={vx,vy,vw,vhCalculating the detection error L by the following formula as a target translation scaling parameterbox
Figure FDA0002402432790000042
wherein
Figure FDA0002402432790000043
For the smoothed norm loss function, tableShown as follows:
Figure FDA0002402432790000044
8. the method of claim 5, wherein (3c7) the binary mask is obtained through a mask branch, and the following is implemented:
(3c7a) feature map F is transformed by deconvolution3Amplifying to obtain binary mask regions M of all categoriesk
(3c7b) traversing all classes of binary mask regions MkFor each class binary mask region MiApplying Sigmoid activation functions
Figure FDA0002402432790000045
After classification operation is carried out, a classification probability vector H is obtained, and the maximum probability y in the H is a binary mask of a target area corresponding to target classification;
(3c7c) calculating the segmentation error L of the mask branchmask
Figure FDA0002402432790000046
wherein ,
Figure FDA0002402432790000047
the true label of the binary mask of the target area.
9. The method of claim 1, wherein the test image is input into the trained model for testing in (4) as follows:
(4a) collecting the test image D obtained in (1a)6Inputting the feature vectors into a trained network model, extracting the feature vectors, obtaining classified probability vectors P by calculating the similarity of the feature vectors, selecting the maximum probability value q in P, and taking the category corresponding to q as a target classification result;
(4b) judging whether the category corresponding to q belongs to a diseased leaf or not:
if the class corresponding to q is a healthy leaf, directly outputting a target detection result and a target classification result;
if the category corresponding to q belongs to the damaged leaf, the q and the damaged leaf receiving threshold T are used0And (3) comparison:
if q > T0Outputting target detection, target classification and segmentation results of disease positions;
if q < ═ T0And directly outputting the target detection result and the target classification result.
10. The method of claim 9, wherein the probability vector of the classification obtained by calculating the similarity of the feature vectors in (4a) is obtained by calculating the cosine similarity cos (θ) between the feature vector a of each real target classification and the feature vector B of the predicted target classification, and the formula is as follows:
Figure FDA0002402432790000051
wherein, | a | | represents the two-norm of the feature vector of the real target classification, and | B | | represents the two-norm of the feature vector of the prediction target classification.
CN202010150980.9A 2020-03-06 2020-03-06 Plant leaf disease identification method based on mask convolutional neural network Active CN111369540B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010150980.9A CN111369540B (en) 2020-03-06 2020-03-06 Plant leaf disease identification method based on mask convolutional neural network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010150980.9A CN111369540B (en) 2020-03-06 2020-03-06 Plant leaf disease identification method based on mask convolutional neural network

Publications (2)

Publication Number Publication Date
CN111369540A true CN111369540A (en) 2020-07-03
CN111369540B CN111369540B (en) 2023-06-02

Family

ID=71210311

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010150980.9A Active CN111369540B (en) 2020-03-06 2020-03-06 Plant leaf disease identification method based on mask convolutional neural network

Country Status (1)

Country Link
CN (1) CN111369540B (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112115888A (en) * 2020-09-22 2020-12-22 四川大学 Plant disease diagnosis system based on disease spot correlation
CN112560644A (en) * 2020-12-11 2021-03-26 四川大学 Crop disease and insect pest automatic identification method suitable for field
CN112598031A (en) * 2020-12-08 2021-04-02 北京农业信息技术研究中心 Vegetable disease detection method and system
CN112634147A (en) * 2020-12-09 2021-04-09 上海健康医学院 PET image noise reduction method, system, device and medium for self-supervision learning
CN112699941A (en) * 2020-12-31 2021-04-23 浙江科技学院 Plant disease severity image classification method and device, computer equipment and storage medium
CN112884022A (en) * 2021-01-29 2021-06-01 浙江师范大学 Unsupervised depth characterization learning method and system based on image translation
CN113052799A (en) * 2021-03-09 2021-06-29 重庆大学 Osteosarcoma and osteochondroma prediction method based on Mask RCNN network
CN113112498A (en) * 2021-05-06 2021-07-13 东北农业大学 Grape leaf scab identification method based on fine-grained countermeasure generation network
CN113191334A (en) * 2021-05-31 2021-07-30 广西师范大学 Plant canopy dense leaf counting method based on improved CenterNet
CN113239788A (en) * 2021-05-11 2021-08-10 嘉兴学院 Mask R-CNN-based wireless communication modulation mode identification method
CN113762190A (en) * 2021-09-15 2021-12-07 中科微至智能制造科技江苏股份有限公司 Neural network-based parcel stacking detection method and device
CN114239756A (en) * 2022-02-25 2022-03-25 科大天工智能装备技术(天津)有限公司 Insect pest detection method and system
CN114742204A (en) * 2022-04-08 2022-07-12 黑龙江惠达科技发展有限公司 Method and device for detecting straw coverage rate
CN114943988A (en) * 2022-06-16 2022-08-26 浙大城市学院 Planar target monitoring method based on instance segmentation and deep convolutional neural network
CN117011718A (en) * 2023-10-08 2023-11-07 之江实验室 Plant leaf fine granularity identification method and system based on multiple loss fusion
CN117372790A (en) * 2023-12-08 2024-01-09 浙江托普云农科技股份有限公司 Plant leaf shape classification method, system and device
CN117882546A (en) * 2024-03-13 2024-04-16 山西诚鼎伟业科技有限责任公司 Intelligent planting method for agricultural operation robot
CN118072177A (en) * 2024-04-12 2024-05-24 国网江西省电力有限公司电力科学研究院 Line channel inflammable tree species identification method based on laser point cloud and image fusion

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190065817A1 (en) * 2017-08-29 2019-02-28 Konica Minolta Laboratory U.S.A., Inc. Method and system for detection and classification of cells using convolutional neural networks
CN110163207A (en) * 2019-05-20 2019-08-23 福建船政交通职业学院 One kind is based on Mask-RCNN ship target localization method and storage equipment
CN110717903A (en) * 2019-09-30 2020-01-21 天津大学 Method for detecting crop diseases by using computer vision technology

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190065817A1 (en) * 2017-08-29 2019-02-28 Konica Minolta Laboratory U.S.A., Inc. Method and system for detection and classification of cells using convolutional neural networks
CN110163207A (en) * 2019-05-20 2019-08-23 福建船政交通职业学院 One kind is based on Mask-RCNN ship target localization method and storage equipment
CN110717903A (en) * 2019-09-30 2020-01-21 天津大学 Method for detecting crop diseases by using computer vision technology

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王艳玲等: "基于迁移学习的番茄叶片病害图像分类", 《中国农业大学学报》 *

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112115888B (en) * 2020-09-22 2022-06-03 四川大学 Plant disease diagnosis system based on disease spot correlation
CN112115888A (en) * 2020-09-22 2020-12-22 四川大学 Plant disease diagnosis system based on disease spot correlation
CN112598031A (en) * 2020-12-08 2021-04-02 北京农业信息技术研究中心 Vegetable disease detection method and system
CN112634147A (en) * 2020-12-09 2021-04-09 上海健康医学院 PET image noise reduction method, system, device and medium for self-supervision learning
CN112634147B (en) * 2020-12-09 2024-03-29 上海健康医学院 PET image noise reduction method, system, device and medium for self-supervision learning
CN112560644A (en) * 2020-12-11 2021-03-26 四川大学 Crop disease and insect pest automatic identification method suitable for field
CN112560644B (en) * 2020-12-11 2021-09-28 四川大学 Crop disease and insect pest automatic identification method suitable for field
CN112699941A (en) * 2020-12-31 2021-04-23 浙江科技学院 Plant disease severity image classification method and device, computer equipment and storage medium
CN112699941B (en) * 2020-12-31 2023-02-14 浙江科技学院 Plant disease severity image classification method, device, equipment and storage medium
CN112884022A (en) * 2021-01-29 2021-06-01 浙江师范大学 Unsupervised depth characterization learning method and system based on image translation
CN113052799A (en) * 2021-03-09 2021-06-29 重庆大学 Osteosarcoma and osteochondroma prediction method based on Mask RCNN network
CN113112498A (en) * 2021-05-06 2021-07-13 东北农业大学 Grape leaf scab identification method based on fine-grained countermeasure generation network
CN113112498B (en) * 2021-05-06 2024-01-19 东北农业大学 Grape leaf spot identification method based on fine-grained countermeasure generation network
CN113239788A (en) * 2021-05-11 2021-08-10 嘉兴学院 Mask R-CNN-based wireless communication modulation mode identification method
CN113191334B (en) * 2021-05-31 2022-07-01 广西师范大学 Plant canopy dense leaf counting method based on improved CenterNet
CN113191334A (en) * 2021-05-31 2021-07-30 广西师范大学 Plant canopy dense leaf counting method based on improved CenterNet
CN113762190A (en) * 2021-09-15 2021-12-07 中科微至智能制造科技江苏股份有限公司 Neural network-based parcel stacking detection method and device
CN113762190B (en) * 2021-09-15 2024-03-29 中科微至科技股份有限公司 Method and device for detecting package stacking based on neural network
CN114239756B (en) * 2022-02-25 2022-05-17 科大天工智能装备技术(天津)有限公司 Insect pest detection method and system
CN114239756A (en) * 2022-02-25 2022-03-25 科大天工智能装备技术(天津)有限公司 Insect pest detection method and system
CN114742204A (en) * 2022-04-08 2022-07-12 黑龙江惠达科技发展有限公司 Method and device for detecting straw coverage rate
CN114943988A (en) * 2022-06-16 2022-08-26 浙大城市学院 Planar target monitoring method based on instance segmentation and deep convolutional neural network
CN114943988B (en) * 2022-06-16 2024-04-02 浙大城市学院 Planar target monitoring method based on instance segmentation and deep convolution neural network
CN117011718B (en) * 2023-10-08 2024-02-02 之江实验室 Plant leaf fine granularity identification method and system based on multiple loss fusion
CN117011718A (en) * 2023-10-08 2023-11-07 之江实验室 Plant leaf fine granularity identification method and system based on multiple loss fusion
CN117372790A (en) * 2023-12-08 2024-01-09 浙江托普云农科技股份有限公司 Plant leaf shape classification method, system and device
CN117372790B (en) * 2023-12-08 2024-03-08 浙江托普云农科技股份有限公司 Plant leaf shape classification method, system and device
CN117882546A (en) * 2024-03-13 2024-04-16 山西诚鼎伟业科技有限责任公司 Intelligent planting method for agricultural operation robot
CN117882546B (en) * 2024-03-13 2024-05-24 山西诚鼎伟业科技有限责任公司 Intelligent planting method for agricultural operation robot
CN118072177A (en) * 2024-04-12 2024-05-24 国网江西省电力有限公司电力科学研究院 Line channel inflammable tree species identification method based on laser point cloud and image fusion

Also Published As

Publication number Publication date
CN111369540B (en) 2023-06-02

Similar Documents

Publication Publication Date Title
CN111369540B (en) Plant leaf disease identification method based on mask convolutional neural network
Dias et al. Multispecies fruit flower detection using a refined semantic segmentation network
WO2022160771A1 (en) Method for classifying hyperspectral images on basis of adaptive multi-scale feature extraction model
CN109344883A (en) Fruit tree diseases and pests recognition methods under a kind of complex background based on empty convolution
CN110222215B (en) Crop pest detection method based on F-SSD-IV3
CN111860330A (en) Apple leaf disease identification method based on multi-feature fusion and convolutional neural network
CN110176005B (en) Remote sensing image segmentation method based on normalized index and multi-scale model
CN110853070A (en) Underwater sea cucumber image segmentation method based on significance and Grabcut
US11037299B2 (en) Region merging image segmentation algorithm based on boundary extraction
CN112749675A (en) Potato disease identification method based on convolutional neural network
Chen et al. Segmentation of field grape bunches via an improved pyramid scene parsing network
Zhang et al. Robust image segmentation method for cotton leaf under natural conditions based on immune algorithm and PCNN algorithm
CN113569772A (en) Remote sensing image farmland instance mask extraction method, system, equipment and storage medium
CN113627481A (en) Multi-model combined unmanned aerial vehicle garbage classification method for smart gardens
CN111832480B (en) Remote sensing identification method for rape planting area based on spectral characteristics
CN113077452A (en) Apple tree pest and disease detection method based on DNN network and spot detection algorithm
Wang et al. MRUNet: A two-stage segmentation model for small insect targets in complex environments
CN114596502B (en) Arable land information extraction method based on edge-enhanced high-resolution remote sensing image
CN113723833B (en) Method, system, terminal equipment and storage medium for evaluating quality of forestation actual results
CN114663791A (en) Branch recognition method for pruning robot in unstructured environment
Rony et al. BottleNet18: Deep Learning-Based Bottle Gourd Leaf Disease Classification
CN115601634A (en) Image blade identification method and device based on hierarchical attention mechanism
Fan et al. Deep neural de-raining model based on dynamic fusion of multiple vision tasks
CN111046861A (en) Method for identifying whether sample equipment exists in infrared image, method for constructing power equipment identification model and application
Song et al. Crop Classification Using Multitemporal Landsat 8 Images

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant