WO2021218215A1 - Image detection method and relevant model training method, relevant apparatuses, and device - Google Patents
Image detection method and relevant model training method, relevant apparatuses, and device Download PDFInfo
- Publication number
- WO2021218215A1 WO2021218215A1 PCT/CN2020/140325 CN2020140325W WO2021218215A1 WO 2021218215 A1 WO2021218215 A1 WO 2021218215A1 CN 2020140325 W CN2020140325 W CN 2020140325W WO 2021218215 A1 WO2021218215 A1 WO 2021218215A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- detection model
- image
- organ
- medical image
- original
- Prior art date
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 754
- 238000012549 training Methods 0.000 title claims abstract description 177
- 238000000034 method Methods 0.000 title claims abstract description 115
- 210000000056 organ Anatomy 0.000 claims abstract description 242
- 230000006870 function Effects 0.000 claims description 22
- 238000012545 processing Methods 0.000 claims description 13
- 238000002372 labelling Methods 0.000 claims description 11
- 238000007781 pre-processing Methods 0.000 claims description 11
- 238000004590 computer program Methods 0.000 claims description 7
- 239000003814 drug Substances 0.000 claims description 6
- 210000003734 kidney Anatomy 0.000 description 30
- 210000000952 spleen Anatomy 0.000 description 30
- 210000000496 pancreas Anatomy 0.000 description 25
- 210000004185 liver Anatomy 0.000 description 16
- 238000002591 computed tomography Methods 0.000 description 11
- 238000010586 diagram Methods 0.000 description 11
- 210000001015 abdomen Anatomy 0.000 description 10
- 230000001186 cumulative effect Effects 0.000 description 10
- 238000004364 calculation method Methods 0.000 description 7
- 238000013527 convolutional neural network Methods 0.000 description 4
- 238000002595 magnetic resonance imaging Methods 0.000 description 4
- 230000003187 abdominal effect Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000013434 data augmentation Methods 0.000 description 3
- 210000002569 neuron Anatomy 0.000 description 3
- 210000000133 brain stem Anatomy 0.000 description 2
- 210000001638 cerebellum Anatomy 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 210000002451 diencephalon Anatomy 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 210000004072 lung Anatomy 0.000 description 2
- 230000001568 sexual effect Effects 0.000 description 2
- 210000001587 telencephalon Anatomy 0.000 description 2
- 210000001685 thyroid gland Anatomy 0.000 description 2
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000004195 computer-aided diagnosis Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 238000002059 diagnostic imaging Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000001959 radiotherapy Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/0002—Inspection of images, e.g. flaw detection
- G06T7/0012—Biomedical image inspection
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/20—Image enhancement or restoration using local operators
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10081—Computed x-ray tomography [CT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10072—Tomographic images
- G06T2207/10088—Magnetic resonance imaging [MRI]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Definitions
- the present disclosure relates to the field of artificial intelligence technology, and in particular to an image detection method and related model training method and related devices and equipment.
- Medical images such as CT (Computed Tomography) and MRI (Magnetic Resonance Imaging, MRI scan) have important clinical significance.
- CT Computer Tomography
- MRI Magnetic Resonance Imaging, MRI scan
- multi-organ detection on medical images such as CT and MRI to determine the region corresponding to each organ on the medical image
- training an image detection model suitable for multi-organ detection has high application value.
- model training relies on a large number of labeled data sets.
- obtaining a large number of high-quality multi-organ annotations is very time-consuming and labor-intensive, and usually only experienced radiologists have the ability to annotate data.
- the existing image detection models often have the problem of low accuracy when performing multi-organ detection. In view of this, how to improve the accuracy of detection in multi-organ detection has become an urgent problem to be solved.
- the present disclosure provides an image detection method, a training method of related models, and related devices and equipment.
- an embodiment of the present disclosure provides a method for training an image detection model, including: obtaining a sample medical image, wherein the sample medical image pseudo-labels at least one actual region of an unlabeled organ; The image is detected to obtain a first detection result, where the first detection result includes the first predicted region of the unlabeled organ; and the sample medical image is detected using the image detection model to obtain the second detection result, where the second detection The results include the second prediction area of the unlabeled organ.
- the network parameters of the image detection model are determined based on the network parameters of the original detection model; the differences between the first prediction area and the actual area and the second prediction area are used to adjust the original detection The network parameters of the model.
- the sample medical image pseudo-labels the actual area of at least one unlabeled organ, there is no need to actually label multiple organs in the sample medical image
- the original detection model is used to detect the sample medical image to obtain Contain the first detection result of the first preset area of the unlabeled organ, and use the image detection model to detect the sample medical image to obtain the second detection result of the second prediction area containing the unlabeled organ, and then use the first prediction area
- the original detection model includes a first original detection model and a second original detection model
- the image detection model includes a first image detection model corresponding to the first original detection model and a second image detection model corresponding to the second original detection model
- Using the original detection model to detect the sample medical image to obtain the first detection result includes: using the first original detection model and the second original detection model to perform the step of detecting the sample medical image to obtain the first detection result; using the image The detection model detects the sample medical image to obtain the second detection result, including: using the first image detection model and the second image detection model to perform the step of detecting the sample medical image to obtain the second detection result; using the first prediction Adjust the network parameters of the original detection model based on the differences between the regions and the actual region and the second prediction region, including: using the first prediction region of the first original detection model to compare the second prediction of the actual region and the second image detection model Adjust the network parameters of the first original detection model; and adjust the difference between the first prediction area of the second original detection model and the actual area and the second prediction area of the
- the original detection model is set to include the first original detection model and the second original detection model
- the image detection model is set to include the first image detection model corresponding to the first original detection model and the image detection model corresponding to the second original detection model.
- the second image detection model, and the first original detection model and the second original detection model are used to perform the step of detecting the sample medical image to obtain the first detection result
- the first image detection model and the second detection model are used respectively to execute
- the step of detecting the sample medical image to obtain the second detection result so as to use the difference between the first prediction area of the first original detection model and the actual area and the second prediction area of the second image detection model to adjust the first
- the network parameters of the original detection model, and the difference between the first prediction area of the second original detection model and the actual area and the second prediction area of the first image detection model are used to adjust the network parameters of the second original detection model.
- the first image detection model corresponding to the first original detection model can be used to supervise the training of the second original detection model
- the second image detection model corresponding to the second original detection model can be used to supervise the training of the first original detection model. Constrain the cumulative error of the network parameters due to the pseudo-labeled real area during multiple training processes, and improve the accuracy of the image detection model.
- using the difference between the first prediction area and the actual area and the second prediction area to adjust the network parameters of the original detection model includes: using the difference between the first prediction area and the actual area to determine the first prediction of the original detection model. Loss value; and, using the difference between the first prediction region and the second prediction region, determine the second loss value of the original detection model; use the first loss value and the second loss value to adjust the network parameters of the original detection model.
- the first loss value of the original detection model is determined by the difference between the first prediction region and the actual region
- the second loss value of the original detection model is determined by the difference between the first prediction region and the second prediction region
- the two dimensions of the difference between the second prediction regions are used to measure the loss of the original detection model, which is helpful to improve the accuracy of loss calculation, which can help improve the accuracy of the network parameters of the original detection model, and thus can help improve the image detection model. Accuracy.
- using the difference between the first prediction area and the actual area to determine the first loss value of the original detection model includes at least one of the following: using a focus loss function to process the first prediction area and the actual area to obtain the first focus loss Value; the first prediction area and the actual area are processed using the ensemble similarity loss function to obtain the first loss value of the ensemble similarity.
- using the difference between the first prediction region and the second prediction region to determine the second loss value of the original detection model includes: using a consistency loss function to process the first prediction region and the second prediction region to obtain the second loss value.
- using the first loss value and the second loss value to adjust the network parameters of the original detection model includes: weighting the first loss value and the second loss value to obtain the weighted loss value; using the weighted loss value to adjust the original detection model Network parameters.
- the model can increase the focus on difficult samples, which can help improve the accuracy of the image detection model;
- the ensemble similarity loss function processes the first prediction area and the actual area to obtain the first loss value of the ensemble similarity, which can make the model fit the pseudo-labeled actual area, which can help improve the accuracy of the image detection model;
- the consistency loss function processes the first prediction area and the second prediction area to obtain the second loss value, which can improve the prediction consistency of the original model and the image detection model, and thus can help improve the accuracy of the image detection model; Perform weighting processing on the first loss value and the second loss value to obtain the weighted loss value, and use the weighted loss value to adjust the network parameters of the original detection model, which can balance the importance of each loss value in the training process, thereby improving the network
- the accuracy of the parameters can help improve the accuracy of the image detection model.
- the sample medical image also contains the actual area of the marked organ
- the first detection result also includes the first prediction area of the marked organ
- the second detection result also includes the second prediction area of the marked organ
- using the first prediction area Determine the first loss value of the original detection model based on the difference between the actual area and the original detection model, including: using the difference between the first prediction area and the actual area of the unlabeled organ and the labeled organ to determine the first loss value of the original detection model
- Using the difference between the first prediction area and the second prediction area to determine the second loss value of the original detection model, including: using the difference between the first prediction area of the unlabeled organ and the corresponding second prediction area to determine the original Check the second loss value of the model.
- the second detection result also includes the second prediction region of the labeled organ
- the difference between the first prediction area and the actual area is comprehensively considered.
- the difference between the prediction area and the corresponding second prediction area can improve the robustness of the consistency constraints of the original detection model and the image detection model, and thus can improve the accuracy of the image detection model.
- the original detection model after adjusting the network parameters of the original detection model by using the differences between the first prediction area and the actual area and the second prediction area, it also includes: using the network parameters adjusted during this training and several previous trainings to correct The network parameters of the image detection model are updated.
- the network parameters can be further constrained due to pseudo-labeled real regions during multiple training sessions.
- the resulting cumulative error improves the accuracy of the image detection model.
- the network parameters adjusted during this training and several previous trainings to update the network parameters of the image detection model, including: statistics the average of the network parameters adjusted by the original detection model during this training and several previous trainings Value; update the network parameters of the image detection model to the average value of the network parameters of the corresponding original detection model.
- acquiring the sample medical image includes: acquiring a medical image to be pseudo-labeled, wherein at least one unlabeled organ exists in the medical image to be pseudo-labeled; and detecting the pseudo-labeled medical image using a single-organ detection model corresponding to each unlabeled organ. , To obtain the organ prediction area of each unlabeled organ; pseudo-label the organ prediction area of the unlabeled organ as the actual area of the unlabeled organ, and use the pseudo-labeled medical image to be pseudo-labeled as the sample medical image.
- the single-organ detection model can be used to avoid the manual labeling of multiple organs. Workload, which can help reduce the labor cost of training an image detection model for multi-organ detection, and improve the efficiency of training.
- the medical image to be pseudo-labeled includes at least one labeled organ; before the single-organ detection model corresponding to each unlabeled organ is used to detect the pseudo-labeled medical image, the method further includes: using the medical image to be pseudo-labeled, Annotate the single-organ detection model corresponding to the annotated organ in the medical image for training.
- the medical image to be pseudo-labeled including at least one labeled organ in the medical image to be pseudo-labeled, and using the medical image to be pseudo-labeled to train the single-organ detection model corresponding to the labeled organ in the pseudo-labeled medical image can improve the accuracy of the single-organ detection model. Therefore, it can help improve the accuracy of subsequent pseudo-labeling, and in turn, can help improve the accuracy of the subsequent training image detection model.
- acquiring a medical image to be pseudo-labeled includes: acquiring a three-dimensional medical image and preprocessing the three-dimensional medical image; and performing cropping processing on the pre-processed three-dimensional medical image to obtain at least one two-dimensional medical image to be pseudo-labeled.
- the pre-processed three-dimensional medical images are cropped to obtain at least one two-dimensional medical image to be pseudo-labeled, which can help to obtain a model training Medical images can help improve the accuracy of subsequent image detection model training.
- the preprocessing of the three-dimensional medical image includes at least one of the following: adjusting the voxel resolution of the three-dimensional medical image to a preset resolution; using a preset window value to normalize the voxel value of the three-dimensional medical image to Within a preset range; Gaussian noise is added to at least part of the voxels of the three-dimensional medical image.
- adjusting the voxel resolution of the 3D medical image to a preset resolution can facilitate subsequent model prediction processing; using the preset window value to normalize the voxel value of the 3D medical image to a preset range can be It is helpful for the model to extract accurate features; adding Gaussian noise to at least part of the voxels of the three-dimensional medical image can help achieve data augmentation, increase data diversity, and improve the accuracy of subsequent model training.
- the embodiments of the present disclosure provide an image detection method, including: acquiring a medical image to be detected, wherein the medical image to be detected contains multiple organs; and using an image detection model to detect the medicine to be detected to obtain multiple organs The prediction area; wherein, the image detection model is obtained by using the training method of the image detection model in the first aspect.
- the detection accuracy can be improved in the process of multiple organ detection.
- an embodiment of the present disclosure provides a training device for an image detection model, including an image acquisition module, a first detection module, a second detection module, and a parameter adjustment module.
- the image acquisition module is configured to acquire sample medical images, wherein , The sample medical image pseudo-labels the actual area of at least one unlabeled organ;
- the first detection module is configured to use the original detection model to detect the sample medical image to obtain the first detection result, wherein the first detection result includes the unlabeled organ
- the second detection module is configured to use the image detection model to detect the sample medical image to obtain the second detection result, and the network parameters of the image detection model are determined based on the network parameters of the original detection model, wherein ,
- the second detection result includes a second prediction area of the unlabeled organ;
- the parameter adjustment module is configured to adjust the network parameters of the original detection model by using the difference between the first prediction area and the actual area and the second prediction area, respectively.
- an embodiment of the present disclosure provides an image detection device, including an image acquisition module and an image detection module, the image acquisition module is configured to acquire a medical image to be detected, wherein the medical image to be detected contains multiple organs; the image The detection module is configured to use the image detection model to detect the medicine to be detected to obtain the predicted regions of multiple organs; wherein the image detection model is obtained by training using the image detection model training device in the second aspect.
- embodiments of the present disclosure provide an electronic device including a memory and a processor coupled to each other.
- the processor is configured to execute program instructions stored in the memory to implement the image detection model in the first aspect. Training method, or implement the image detection method in the second aspect.
- embodiments of the present disclosure provide a computer-readable storage medium on which program instructions are stored.
- the program instructions are executed by a processor, the training method of the image detection model in the first aspect is realized, or the first aspect is realized.
- the image detection method in the second aspect is realized.
- the embodiments of the present disclosure also provide a computer program, including computer-readable code.
- the processor in the electronic device executes the above-mentioned first aspect.
- the sample medical image is acquired, and the sample medical image is pseudo-labeled with at least one actual region of an unlabeled organ, so there is no need to actually label multiple organs in the sample medical image, and the original detection model is used to detect the sample medical image.
- the difference between the area and the actual area and the second predicted area, adjust the network parameters of the original detection model, and the network parameters of the image detection model are determined based on the network parameters of the original detection model, so the image detection model can supervise the original detection
- the training of the model can constrain the cumulative error of the network parameters due to the pseudo-labeled real area during multiple training processes, and improve the accuracy of the image detection model, so that the image detection model can accurately supervise the training of the original detection model.
- the original detection model can accurately adjust its network parameters during the training process. Therefore, the detection accuracy of the image detection model can be improved in the process of multi-organ detection.
- FIG. 1 is a schematic flowchart of an embodiment of a training method for an image detection model provided by an embodiment of the present disclosure
- FIG. 2 is a schematic flowchart of an embodiment of step S11 in FIG. 1;
- FIG. 3 is a schematic flowchart of another embodiment of a training method for an image detection model provided by an embodiment of the present disclosure
- FIG. 4 is a schematic diagram of an embodiment of the training process of an image detection model provided by an embodiment of the present disclosure
- FIG. 5 is a schematic flowchart of an embodiment of an image detection method provided by an embodiment of the present disclosure
- FIG. 6 is a schematic diagram of the framework of an embodiment of an image detection model training apparatus provided by an embodiment of the present disclosure
- FIG. 7 is a schematic diagram of a framework of an embodiment of an image detection device provided by an embodiment of the present disclosure.
- FIG. 8 is a schematic diagram of a framework of an embodiment of an electronic device provided by an embodiment of the present disclosure.
- FIG. 9 is a schematic framework diagram of an embodiment of a computer-readable storage medium provided by an embodiment of the present disclosure.
- system and "network” in this article are often used interchangeably in this article.
- the term “and/or” in this article is only an association relationship describing the associated objects, which means that there can be three relationships, for example, A and/or B, which can mean: A alone exists, A and B exist at the same time, exist alone B these three situations.
- the character "/” in this text generally indicates that the associated objects before and after are in an "or” relationship.
- "many” in this document means two or more than two.
- FIG. 1 is a schematic flowchart of an embodiment of a method for training an image detection model provided by an embodiment of the present disclosure. Among them, the following steps can be included:
- Step S11 Obtain a sample medical image, where the sample medical image pseudo-labels at least one actual region of an unlabeled organ.
- the sample medical images may include CT images and MR images, which are not limited here.
- the sample medical image can be obtained by scanning the abdomen, chest, head, etc., and can be set according to actual application conditions, which is not limited here.
- the organs in the sample medical image may include: kidney, spleen, liver, pancreas, etc.; or, when the chest is scanned, the organs in the sample medical image may include: heart, lung lobes, thyroid, etc.; or,
- the head is scanned, and the organs in the sample medical image can include: brain stem, cerebellum, diencephalon, and telencephalon.
- the actual area of the unlabeled organ may be detected by using a single-organ detection model corresponding to the unlabeled organ.
- the unlabeled organ may include: At least one of the kidney, spleen, liver, and pancreas can use the single organ detection model corresponding to the kidney to detect the sample medical image to obtain the organ prediction area corresponding to the kidney, and the single organ detection model corresponding to the spleen can be used Detect the sample medical image to obtain the organ prediction area corresponding to the spleen, and use the single organ detection model corresponding to the liver to detect the sample medical image to obtain the organ prediction area corresponding to the liver, and use the single organ detection corresponding to the pancreas
- the model detects the sample medical image and obtains the organ prediction region corresponding to the pancreas, so that the organ prediction regions corresponding to the kidney, spleen, liver, and pancreas are pseudo-labeled in the sample medical image,
- pseudo-labeling refers to the process of taking the organ prediction regions of unlabeled organs detected by the single-organ detection model as the actual regions.
- the organ is not marked as other organs, it can be deduced by analogy, and we will not give examples one by one here.
- the single-organ detection model for unlabeled organs is trained using a single-organ data set labeled with the actual region of the unlabeled organ.
- the single-organ detection model corresponding to the kidney uses the labeled kidney
- the kidney data set of the actual area is trained, and the single-organ detection model corresponding to the spleen is trained using the spleen data set of the actual area marked with the spleen.
- Step S12 Use the original detection model to detect the sample medical image to obtain a first detection result, where the first detection result includes a first prediction region of an unlabeled organ.
- the original detection model can include any one of Mask R-CNN (Mask Region with Convolutional Neural Network), FCN (Fully Convolutional Network), PSP-net (Pyramid Scene Parsing Network, pyramid scene analysis network),
- the original detection model can also be set-net, U-net, etc., which can be set according to the actual situation, which is not limited here.
- the first detection result of the first prediction region containing the unlabeled organ can be obtained.
- the sample medical image is an image obtained by scanning the abdomen.
- the unlabeled organs include the kidney, spleen, and pancreas. Therefore, the original detection model is used to detect the sample medical image, and the first prediction area of the kidney and the first prediction area of the spleen can be obtained.
- the first prediction area of the pancreas, and other scenarios can be deduced by analogy, so I won’t give an example one by one here.
- Step S13 Use the image detection model to detect the sample medical image to obtain a second detection result, where the second detection result includes a second prediction region of an unlabeled organ.
- the network structure of the original detection model and the network structure of the image detection model corresponding to the original detection model may be the same.
- the corresponding image detection model can also be Mask R-CNN; or, when the original detection model is FCN, the corresponding image detection model can also be FCN; Or, when the original detection model is PSP-net, the corresponding image detection model can also be PSP-net; when the original detection model is another network, the analogy can be used, and no examples are given here.
- the network parameters of the image detection model may be determined based on the network parameters of the original detection model.
- the network parameters of the image detection model may be obtained based on the network parameters adjusted by the original detection model in multiple training processes.
- the network parameters of the image detection model can be obtained by using the network parameters adjusted by the original detection model from the knth to the k-1th training process; or, in the k+th
- the network parameters of the image detection model can be obtained by using the network parameters adjusted by the original detection model from the k+1-nth to the kth training process, and so on.
- the number of times (ie, n) of the foregoing multiple trainings can be set according to actual conditions, for example, it can be set to 5, 10, 15, etc., which are not limited here.
- the second detection result of the second prediction region containing the unlabeled organ can be obtained.
- the unlabeled organs include the kidney, spleen, and pancreas. Therefore, the image detection model is used to detect the sample medical image, and the second prediction area of the kidney and the second prediction area of the spleen can be obtained.
- the prediction area, the second prediction area of the pancreas, and other scenarios can be deduced by analogy, so we will not give examples one by one here.
- the above steps S12 and S13 may be performed in a sequential order, for example, step S12 is performed first, and then step S13; or, step S13 is performed first, and then step S12 is performed.
- the above step S12 and step S13 can also be performed at the same time, and can be set according to actual applications, which is not limited here.
- Step S14 Use the differences between the first prediction area and the actual area and the second prediction area to adjust the network parameters of the original detection model.
- the difference between the first prediction area and the actual area can be used to determine the first loss value of the original detection model.
- the focal loss function can be used to process the first prediction area and the actual area to obtain the first focal loss value; or, in order to be able to make the model fit pseudo-labeled In the actual area, the first prediction area and the actual area can also be processed by using the dice loss function to obtain the first loss value of the dice loss.
- the difference between the first prediction area and the second prediction area can also be used to determine the second loss value of the original detection model.
- the consistency loss function can be used to process the first prediction area and the second prediction area to obtain the second loss value.
- the performance loss function can be a cross-entropy loss function, which can be set according to actual application conditions, and is not limited here.
- the above-mentioned first loss value and second loss value can also be used to adjust the network parameters of the original detection model.
- the first loss value and the second loss value can be weighted to obtain a weighted loss value, so that the weighted loss value can be used to adjust the network parameters of the original detection model.
- the weights corresponding to the first loss value and the second loss value can be set according to the actual situation, for example, both are set to 0.5; or, the weight corresponding to the first loss value is set to 0.6, and the weight corresponding to the second loss value is set Set to 0.4, which is not limited here.
- the first loss value includes the first loss value of the focus and the first loss value of the set similarity
- the first loss value of the focus, the first loss value of the set similarity, and the second loss value can be weighted to obtain The weighted loss value is used to adjust the network parameters of the original detection model.
- Stochastic Gradient Descent (SGD), Batch Gradient Descent (BGD), Mini-Batch Gradient Descent (MBGD), etc. can be used, and weighted The loss value adjusts the network parameters of the original detection model.
- batch gradient descent refers to the use of all samples for parameter updates during each iteration; stochastic gradient descent refers to the use of one during each iteration Samples are used to update parameters; mini-batch gradient descent refers to using a batch of samples to update parameters during each iteration, which will not be repeated here.
- the sample medical image may also include the actual area of the marked organ
- the first detection result may also include the first prediction area of the marked organ
- the second detection result may also include the second area of the marked organ. Forecast area.
- the unlabeled organs include the kidney, spleen, and pancreas
- the labeled organs include the liver. Therefore, the original detection model is used to detect the sample medical image, and the corresponding kidneys of the unlabeled organs can be obtained.
- the first prediction area corresponding to the unlabeled organ spleen, the first prediction area corresponding to the unlabeled organ pancreas, and the first prediction area corresponding to the labeled organ liver, and the image detection model corresponding to the original detection model is used Detecting the sample medical image can obtain the second prediction area corresponding to the unlabeled organ kidney, the second prediction area corresponding to the unlabeled organ spleen, the second prediction area corresponding to the unlabeled organ pancreas, and the second prediction area corresponding to the labeled organ liver. Forecast area.
- the difference between the first prediction region and the actual region of the unlabeled organ and the labeled organ can be used to determine the first loss value of the original detection model, and the difference between the first prediction region of the unlabeled organ and the corresponding second prediction region can be used.
- the difference between the two can determine the second loss value of the original detection model.
- the unlabeled organs include the kidney, spleen, and pancreas
- the labeled organs include the liver. You can use the first prediction area corresponding to the unlabeled organ kidney and the pseudo-labeled actual area.
- the difference between the first prediction area corresponding to the unlabeled organ spleen and the pseudo-labeled actual area, the difference between the first prediction area corresponding to the unlabeled organ pancreas and the pseudo-labeled actual area, and the labeled organ liver Determine the first loss value of the original detection model according to the difference between the corresponding first prediction area and the actual area marked by the real label.
- the first loss value may include at least one of the first loss value of focus and the first loss value of set similarity. Alternatively, please refer to the previous steps, which will not be repeated here.
- the difference between the first prediction region and the second prediction region corresponding to the unlabeled organ kidney, the difference between the first prediction region and the second prediction region corresponding to the spleen of the unlabeled organ, and the pancreas corresponding to the unlabeled organ is determined to determine the second loss value of the original detection model.
- the second loss value can be calculated by using the cross-entropy loss function. You can refer to the foregoing steps and will not be repeated here. Therefore, in the process of determining the first loss value of the original detection model, the difference between the first prediction area and the actual area is comprehensively considered, and in the process of determining the second loss value of the original detection model, only unlabeled organs are considered.
- the difference between the first prediction region and the corresponding second prediction region can improve the robustness of the consistency constraint of the original detection model and the image detection model, and thus can improve the accuracy of the image detection model.
- the network parameters of the image detection model may not be updated, but after a preset number of times (for example, 2 times, 3 times, etc.) training, reuse The network parameters adjusted during this training and several previous trainings will update the network parameters of the image detection model, which is not limited here. For example, during the kth training process, the network parameters of the image detection model may not be updated.
- the original detection model can be used to train from the k+inth to the k+ith time.
- i can be set to an integer not less than 1 according to the actual situation, for example, it can be set to 1, 2, 3, etc., which is not limited here.
- the original detection model in the process of updating the network parameters of the image detection model, can be counted in this training and the average value of the network parameters adjusted by several previous trainings, and then the image detection model The network parameters of is updated to the average value of the network parameters of the corresponding original detection model.
- the average value of network parameters refers to the average value corresponding to the same network parameter, which may be a certain weight (or bias) corresponding to the same neuron after being adjusted in multiple training processes.
- the average value of the value, so the average value of each weight (or bias) of each neuron after adjustment in multiple training processes can be obtained by statistics, so as to use the average value to the corresponding weight of the corresponding neuron in the image detection model (Or offset) to update.
- this training is the kth training, and the average value of the network parameters adjusted by the original detection model during this training and the previous n-1 training can be counted.
- the value of n can be set according to the actual application, for example , Can be set to 5, 10, 15, etc., which is not limited here.
- the network parameters of the image detection model are updated using the average value of the adjusted network parameters from the k-n+1 training process to the k training process, so as to be able to It is conducive to quickly constrain the accumulated errors generated in the process of multiple training, and improve the accuracy of the image detection model.
- a preset training end condition can also be set. If the preset training end condition is not met, the above step S12 and subsequent steps can be re-executed to continue to perform the network parameters of the original detection model. Adjustment.
- the preset training end conditions may include any of the following: the current number of training times reaches a preset number threshold (eg, 500 times, 1000 times, etc.), and the loss value of the original detection model is less than a preset loss threshold. For one, there is no limitation here.
- the image detection model can be used to detect the medical image to be tested, so that the regions corresponding to multiple organs in the medical image to be tested can be directly obtained, thereby eliminating the need to use multiple units.
- Organ detection performs separate detection operations on medical images to be detected, so the amount of detection calculations can be reduced.
- the sample medical image is acquired, and the sample medical image is pseudo-labeled with at least one actual region of an unlabeled organ, so there is no need to actually label multiple organs in the sample medical image, and the original detection model is used to detect the sample medical image.
- the difference between the area and the actual area and the second predicted area, adjust the network parameters of the original detection model, and the network parameters of the image detection model are determined by the network parameters of the original detection model, so the image detection model can supervise the original detection
- the training of the model can constrain the cumulative error of the network parameters due to the pseudo-labeled real area during multiple training processes, and improve the accuracy of the image detection model, so that the image detection model can accurately supervise the training of the original detection model.
- the original detection model can accurately adjust its network parameters during the training process. Therefore, the detection accuracy of the image detection model can be improved in the process of multi-organ detection.
- FIG. 2 is a schematic flowchart of an embodiment of step S11 in FIG. 1.
- FIG. 2 is a schematic diagram of an embodiment of obtaining a sample medical image, which includes the following steps:
- Step S111 Obtain a medical image to be pseudo-labeled, where at least one unlabeled organ exists in the medical image to be pseudo-labeled.
- the medical image to be pseudo-labeled can be obtained by scanning the abdomen, the unlabeled organs in the medical image to be pseudo-labeled can include: kidney, spleen, pancreas, etc., and the medical image to be pseudo-labeled can also be obtained by scanning other parts.
- the chest, head, etc. can refer to the relevant steps in the foregoing embodiment, which is not limited here.
- the acquired original medical image can be a three-dimensional medical image, for example, a three-dimensional CT image, a three-dimensional MR image, which is not limited here, so the three-dimensional medical image can be preprocessed and the preprocessed
- the three-dimensional medical image is cropped to obtain at least one medical image to be pseudo-labeled.
- the cropping process may be center cropping of the preprocessed three-dimensional medical image, which is not limited here.
- cropping can be performed along a plane parallel to the three-dimensional medical image in the dimensions perpendicular to the plane to obtain a two-dimensional medical image to be pseudo-labeled.
- the size of the medical image to be pseudo-labeled can be set according to the actual situation, for example, it can be 352*352, which is not limited here.
- the preprocessing may include adjusting the voxel resolution of the three-dimensional medical image to a preset resolution.
- the voxel of the 3D medical image is the smallest unit of 3D medical image segmentation in the 3D space.
- the preset resolution can be 1*1*3mm, and the preset resolution can also be set to other resolutions according to the actual situation, for example, 1*1 *4mm, 2*2*3mm, etc., are not limited here. Adjusting the voxel resolution of the three-dimensional medical image to a preset resolution can facilitate subsequent model prediction processing.
- the preprocessing may also include using a preset window value to normalize the voxel value of the three-dimensional medical image to a preset range.
- the voxel value can be a different value depending on the three-dimensional medical image.
- the voxel value can be a Hu (houns field unit) value.
- the preset window value can be set according to the part corresponding to the 3D medical image.
- the preset window value can be set from -125 to 275, and other parts can be set according to the actual situation. , I will not give an example one by one here.
- the preset range can be set according to the actual application.
- the preset range can be set from 0 to 1, still taking 3D CT images as an example.
- the preset window value can be set from -125 to 275.
- voxels with a voxel value less than or equal to -125 can be reset to a voxel value
- voxels with a voxel value greater than or equal to 275 can be reset to a voxel uniformly.
- Voxel value 1 you can reset voxels with voxel values between -125 to 275 to voxel values between 0 and 1, which can help enhance the contrast between different organs in the image, thereby improving the accuracy of the model extraction feature.
- the preprocessing may also include adding Gaussian noise to at least part of the voxels of the three-dimensional medical image. At least part of the voxels can be set according to actual applications, for example, 1/3 voxels of 3D medical images, or 1/2 voxels of 3D medical images, or all voxels of 3D medical images, which are not limited here. .
- Gaussian noise By adding Gaussian noise to at least part of the voxels of the three-dimensional medical image, the subsequent two-dimensional medical image to be pseudo-labeled can be cropped on the basis of the three-dimensional medical image and the three-dimensional medical image without Gaussian noise, so it can be beneficial to implementation Data augmentation, increase data diversity, and improve the accuracy of subsequent model training.
- Step S112 Use the single-organ detection model corresponding to each unlabeled organ to detect the pseudo-labeled medical image to obtain the organ prediction area of each unlabeled organ.
- the single-organ detection model corresponding to each unlabeled organ may be trained using a single-organ data set labeled with unlabeled organs.
- the single-organ detection model corresponding to the kidney may be obtained by using a single-organ detection model labeled with kidneys.
- the single-organ detection model corresponding to the spleen can be trained using the single-organ data set labeled with the spleen, and the other organs can be deduced by analogy, so we will not give an example one by one here.
- the medical image to be pseudo-labeled may also include at least one labeled organ, and the medical image to be pseudo-labeled including the labeled organ may be used to treat a single organ corresponding to the labeled organ in the pseudo-labeled medical image.
- the detection model is trained to obtain the corresponding single-organ detection model. For example, if the medical image to be pseudo-labeled includes the labeled liver, the medical image to be pseudo-labeled that includes the labeled liver can be used to train the single-organ detection model corresponding to the liver to obtain the single-organ detection model corresponding to the liver. It can be deduced by analogy, and I will not give examples one by one here.
- single-organ detection models can include any of Mask R-CNN (Mask Region with Convolutional Neural Network), FCN (Fully Convolutional Network), and PSP-net (Pyramid Scene Parsing Network).
- Mask R-CNN Mask Region with Convolutional Neural Network
- FCN Full Convolutional Network
- PSP-net Pyramid Scene Parsing Network
- the single-organ detection model can also be set-net, U-net, etc., which can be set according to actual conditions, which is not limited here.
- the organ prediction area of each unlabeled organ can be obtained.
- the medical image to be pseudo-labeled is an image obtained by scanning the abdomen as an example
- the unlabeled organs include the kidney, spleen, and pancreas.
- the single-organ detection model corresponding to the kidney is used to detect the pseudo-labeled medical image, and the organ prediction area of the kidney can be obtained.
- the single-organ detection model corresponding to the kidney can be used to detect pseudo-labeled medical images
- the single-organ detection model corresponding to the spleen can be used to detect pseudo-labeled medical images
- the use of The single-organ detection model corresponding to the pancreas requires the steps of detecting and detecting pseudo-labeled medical images, and finally uniformly pseudo-labeling the single-organ prediction regions of the kidney, spleen, and pancreas on the medical image to be pseudo-labeled; or, using each of the above
- the single-organ detection model corresponding to the unlabeled organ can also perform the steps of detecting the pseudo-labeled medical image in sequence, so that it is no longer necessary to pseudo-label the organ prediction region of each unlabeled organ in the pseudo-labeled medical image.
- the final medical image to be pseudo-labeled can include the single-organ prediction regions of the kidney, spleen, and pancreas. It can be set according to the actual situation and is not limited here.
- Step S113 pseudo-label the organ prediction region of the unlabeled organ as the actual region of the unlabeled organ, and use the pseudo-labeled medical image to be pseudo-labeled as a sample medical image.
- the organ prediction region of each unlabeled organ can be pseudo-labeled as the actual region of the unlabeled organ, and the pseudo-labeled medical image to be pseudo-labeled can be used as the sample medical image.
- the organ prediction area of the non-labeled organ is pseudo-labeled as the actual area of the unlabeled organ, and the pseudo-labeled medical image after the pseudo-labeling is used as the sample medical image.
- the single-organ detection model can be used to eliminate the need for manual pairing. The workload of organ labeling can help reduce the labor cost of training an image detection model for multi-organ detection and improve the efficiency of training.
- FIG. 3 is a schematic flowchart of another embodiment of a training method for an image detection model provided by an embodiment of the present disclosure. Among them, the following steps can be included:
- Step S31 Obtain a sample medical image, where the sample medical image pseudo-labels at least one actual region of an unlabeled organ.
- step S31 can refer to related steps in the foregoing embodiment.
- Step S32 using the first original detection model and the second original detection model to perform the step of detecting the sample medical image to obtain the first detection result.
- the original detection model may include a first original detection model and a second original detection model.
- the first original detection model can include any of Mask R-CNN (Mask Region with Convolutional Neural Network), FCN (Fully Convolutional Network), PSP-net (Pyramid Scene Parsing Network, pyramid scene analysis network)
- the first original detection model can also be set-net, U-net, etc., which can be set according to the actual situation, which is not limited here.
- the second original detection model can include any of Mask R-CNN (Mask Region with Convolutional Neural Network), FCN (Fully Convolutional Network), PSP-net (Pyramid Scene Parsing Network, pyramid scene analysis network)
- the second original detection model can also be set-net, U-net, etc., which can be set according to the actual situation, which is not limited here.
- the first detection result detected by the first original detection model may include the first prediction area of the unlabeled organ, or the first detection result detected by the first original detection model may also include the unlabeled organ. The first prediction area and the first prediction area of the labeled organ.
- the first detection result detected by the second original detection model may include the first prediction region of the unlabeled organ, or the first detection result detected by the second original detection model may also include the unlabeled organ The first prediction area of and the first prediction area of the labeled organ.
- FIG. 4 is a schematic diagram of an embodiment of the training process of the image detection model.
- the first original detection model is denoted as net1
- the second original detection model is denoted as net2.
- the first original detection model net1 detects the sample medical image, and the first detection result corresponding to the first original detection model net1 is obtained.
- the second original detection model net2 detects the sample medical image, and obtains the first detection result corresponding to the first original detection model net1. 2.
- Step S33 using the first image detection model and the second image detection model to perform the step of detecting the sample medical image to obtain the second detection result.
- the image detection model may include a first image detection model corresponding to the first original detection model and a second image detection model corresponding to the second original detection model, the network structure and network parameters of the first image detection model and the second image detection model You can refer to the relevant steps in the foregoing embodiment, which will not be repeated here.
- the second detection result detected by the first image detection model may include the second prediction area of the unlabeled organ, or the second detection result detected by the first image detection model may also include the unlabeled organ.
- the second detection result detected by the second image detection model may include the second prediction area of the unlabeled organ, or the second detection result detected by the second image detection model may also include the unlabeled organ The second prediction area of and the second prediction area of the labeled organ.
- the first image detection model corresponding to the first original detection model net1 is denoted as EMA net1
- the second image detection model corresponding to the second original detection model net2 is denoted as EMAnet2.
- the first image detection model EMAnet1 detects the sample medical image
- the second detection result corresponding to the first image detection model EMAnet1 is obtained
- the second image detection model EMAnet2 detects the sample medical image.
- steps S32 and S33 can be performed in a sequential order, for example, step S32 is performed first, and then step S33 is performed, or step S33 is performed first, and then step S32 is performed.
- the above step S32 and step S33 can also be performed at the same time, and can be set according to actual applications, which is not limited here.
- Step S34 Use the differences between the first prediction area of the first original detection model and the actual area and the second prediction area of the second image detection model to adjust the network parameters of the first original detection model.
- the difference between the first prediction area of the first original detection model and the pseudo-labeled actual area can be used to determine the first loss value of the first original detection model, and the first prediction area of the first original detection model and the pseudo-labeled actual area can be used.
- the difference between the second prediction regions of the second image detection model determines the second loss value of the first original detection model, so that the first loss value and the second loss value are used to adjust the network parameters of the first original detection model.
- the calculation methods of the first loss value and the second loss value can refer to the relevant steps in the foregoing embodiment, and will not be repeated here.
- the process of calculating the second loss value only the first prediction area and the second prediction area of the unlabeled organs can be calculated, so as to improve the consistency between the first original detection model and the second image detection model.
- the robustness of sexual constraints can in turn improve the accuracy of the image detection model.
- Step S35 Use the difference between the first prediction area of the second original detection model and the actual area and the second prediction area of the first image detection model to adjust the network parameters of the second original detection model.
- the difference between the first prediction area of the second original detection model and the pseudo-labeled actual area can be used to determine the first loss value of the second original detection model, and the first prediction area and the pseudo-labeled actual area of the second original detection model can be used.
- the difference between the second prediction regions of the first image detection model determines the second loss value of the second original detection model, so that the first loss value and the second loss value are used to adjust the network parameters of the second original detection model.
- the calculation methods of the first loss value and the second loss value can refer to the relevant steps in the foregoing embodiment, and will not be repeated here.
- only the first prediction area and the second prediction area of the unlabeled organ can be calculated, so as to improve the consistency between the second original detection model and the first image detection model.
- the robustness of sexual constraints can in turn improve the accuracy of the image detection model.
- steps S34 and S35 may be performed in a sequential order, for example, step S34 is performed first, and then step S35 is performed, or step S35 is performed first, and then step S34 is performed.
- the above step S24 and step S35 can also be performed at the same time, and can be set according to actual applications, which is not limited here.
- Step S36 Utilize the network parameters adjusted during the current training of the first original detection model and several previous trainings to update the network parameters of the first image detection model.
- the average value of the network parameters adjusted by the first original detection model during this training and several previous trainings can be counted, and the network parameters of the first image detection model can be updated to the corresponding network parameters of the first original detection model. average value.
- Step S37 The network parameters of the second image detection model are updated by using the network parameters adjusted during the current training of the second original detection model and several previous trainings.
- the average value of the network parameters adjusted by the second original detection model during this training and several previous trainings can be counted, and the network parameters of the second image detection model can be updated to the corresponding network parameters of the second original detection model. average value.
- steps S36 and S37 can be performed in a sequential order, for example, step S36 is performed first, and then step S37, or step S37 is performed first, and step S36 is performed later.
- the above step S36 and step S37 can also be performed at the same time, and can be set according to actual applications, which is not limited here.
- the above step S32 and subsequent steps can be re-executed to continue Adjust the network parameters of the first original detection model and the second original detection model, and adjust the network parameters of the first image detection model corresponding to the first original detection model and the second image detection model corresponding to the second original detection model
- the network parameters are updated.
- the preset training end conditions may include: the current number of training times reaches the preset number threshold (eg, 500 times, 1000 times, etc.), and the loss values of the first original detection model and the second original detection model are less than Any one of a preset loss threshold is not limited here.
- any one of the first image detection model and the second image detection model can be used as the network model for subsequent image detection, so that the number of medical images to be detected can be directly obtained.
- the area corresponding to each organ can eliminate the need to use multiple single organs to detect the medical image to be detected separately, so the amount of detection calculation can be reduced.
- the original detection model is set to include the first original detection model and the second original detection model
- the image detection model is set to include the first image detection model corresponding to the first original detection model and the second original detection model.
- Detect the second image detection model corresponding to the model and use the first original detection model and the second original detection model to perform the step of detecting the sample medical image to obtain the first detection result, and use the first image detection model and the first image detection model respectively.
- the second detection model executes the step of detecting the sample medical image to obtain the second detection result, thereby using the difference between the first prediction area of the first original detection model and the actual area and the second prediction area of the second image detection model.
- Adjust the network parameters of the first original detection model and use the difference between the first prediction area of the second original detection model and the actual area and the second prediction area of the first image detection model to adjust the second original detection model Network parameters, so the first image detection model corresponding to the first original detection model can be used to supervise the training of the second original detection model, and the second image detection model corresponding to the second original detection model can be used to supervise the training of the first original detection model. Therefore, it is possible to further constrain the cumulative error of the network parameters due to the pseudo-labeled real region during multiple training processes, and improve the accuracy of the image detection model.
- FIG. 5 is a schematic flowchart of an embodiment of an image detection method provided by an embodiment of the present disclosure. Among them, the following steps can be included:
- Step S51 Obtain a medical image to be tested, where the medical image to be tested contains multiple organs.
- the medical images to be detected may include CT images and MR images, which are not limited here.
- the medical image to be detected can be obtained by scanning the abdomen, chest, head, etc., and can be set according to actual application conditions, which is not limited here.
- the organs in the medical image to be tested may include: kidney, spleen, liver, pancreas, etc.; or scanning the chest, the organs in the medical image to be tested may include: heart, lung lobes, thyroid, etc.;
- the head is scanned, and the organs in the medical image to be detected may include: brain stem, cerebellum, diencephalon, and telencephalon.
- Step S52 Use the image detection model to detect the medicine to be detected to obtain predicted regions of multiple organs.
- the image detection model is obtained by training using the steps in any of the above-mentioned image detection model training method embodiments. You can refer to the relevant steps in the foregoing embodiment, which will not be repeated here.
- the image detection model to detect the medical image to be detected, the predicted regions of multiple organs can be directly obtained, and the operation of using multiple single organs to detect the medical image to be detected can be avoided, and the amount of detection calculation can be reduced.
- the image detection model trained by using the steps in the embodiment of the training method of any of the above-mentioned image detection models detects and detects the medical image to be detected, and obtains the predicted regions of multiple organs, which can improve the detection in the process of multiple organ detection. accuracy.
- FIG. 6 is a schematic diagram of an embodiment of an image detection model training apparatus provided by an embodiment of the present disclosure.
- the training device 60 for the image detection model includes an image acquisition module 61, a first detection module 62, a second detection module 63, and a parameter adjustment module 64.
- the image acquisition module 61 is configured to acquire sample medical images, wherein the sample medical images are pseudo-labeled At least one actual region of an unlabeled organ; the first detection module 62 is configured to use the original detection model to detect the sample medical image to obtain a first detection result, where the first detection result includes the first predicted region of the unlabeled organ; And, the second detection module 63 is configured to use the image detection model to detect the sample medical image to obtain a second detection result, wherein the second detection result includes a second predicted region of an unlabeled organ, and the network parameter of the image detection model is Determined by using the network parameters of the original detection model; the parameter adjustment module 64 is configured to adjust the network parameters of the original detection model by using the differences between the first prediction area and the actual area and the second prediction area, respectively.
- the sample medical image is acquired, and the sample medical image is pseudo-labeled with at least one actual region of an unlabeled organ, so there is no need to actually label multiple organs in the sample medical image, and the original detection model is used to detect the sample medical image.
- the difference between the area and the actual area and the second predicted area, adjust the network parameters of the original detection model, and the network parameters of the image detection model are determined by the network parameters of the original detection model, so the image detection model can supervise the original detection
- the training of the model can constrain the cumulative error of the network parameters due to the pseudo-labeled real area during multiple training processes, and improve the accuracy of the image detection model, so that the image detection model can accurately supervise the training of the original detection model.
- the original detection model can accurately adjust its network parameters during the training process. Therefore, the detection accuracy of the image detection model can be improved in the process of multi-organ detection.
- the original detection model includes a first original detection model and a second original detection model
- the image detection model includes a first image detection model corresponding to the first original detection model and a second image detection model corresponding to the second original detection model.
- Image detection model the first detection module 62 is also configured to use the first original detection model and the second original detection model to perform the step of detecting the sample medical image to obtain the first detection result
- the second detection model 63 is also configured
- the parameter adjustment module 64 is further configured to use the first prediction area of the first original detection model respectively.
- the difference between the actual area and the second prediction area of the second image detection model is adjusted to adjust the network parameters of the first original detection model.
- the parameter adjustment module 64 is also configured to use the first prediction area of the second original detection model. The difference between the actual area and the second prediction area of the first image detection model is adjusted to adjust the network parameters of the second original detection model.
- the original detection model is set to include the first original detection model and the second original detection model
- the image detection model is set to include the first image detection model corresponding to the first original detection model and the second original detection model.
- Detect the second image detection model corresponding to the model and use the first original detection model and the second original detection model to perform the step of detecting the sample medical image to obtain the first detection result, and use the first image detection model and the first image detection model respectively.
- the second detection model performs the step of detecting the sample medical image to obtain the second detection result, thereby using the difference between the first prediction area of the first original detection model and the actual area and the second prediction area of the second image detection model.
- Adjust the network parameters of the first original detection model and use the difference between the first prediction area of the second original detection model and the actual area and the second prediction area of the first image detection model to adjust the second original detection model Network parameters, so the first image detection model corresponding to the first original detection model can be used to supervise the training of the second original detection model, and the second image detection model corresponding to the second original detection model can be used to supervise the training of the first original detection model. Therefore, it is possible to further constrain the cumulative error of the network parameters due to the pseudo-labeled real area during multiple training processes, and improve the accuracy of the image detection model.
- the parameter adjustment module 64 includes a first loss determination sub-module configured to use the difference between the first prediction area and the actual area to determine the first loss value of the original detection model, and the parameter adjustment module 64 includes a first loss value.
- the second loss determination sub-module is configured to use the difference between the first prediction area and the second prediction area to determine the second loss value of the original detection model.
- the parameter adjustment module 64 includes a parameter adjustment sub-module configured to use the first The loss value and the second loss value adjust the network parameters of the original detection model.
- the first loss value of the original detection model is determined by the difference between the first prediction area and the actual area, and the difference between the first prediction area and the second prediction area is used to determine the value of the original detection model.
- the second loss value and use the first loss value and the second loss value to adjust the network parameters of the original detection model, so that the first prediction area predicted from the original detection model can be detected with the pseudo-labeled actual area and the corresponding image respectively.
- the two dimensions of the difference between the second prediction regions predicted by the model are used to measure the loss of the original detection model, which is conducive to improving the accuracy of the loss calculation, which can help improve the accuracy of the network parameters of the original detection model, which in turn can help Improve the accuracy of the image detection model.
- the first loss determination submodule includes a focus loss determination unit configured to process the first prediction area and the actual area using a focus loss function to obtain the first focus loss value
- the first loss determination submodule includes The collective similarity loss determination unit is configured to use the collective similarity loss function to process the first prediction region and the actual region to obtain the first loss value of the collective similarity
- the second loss determination sub-module is also configured to use the consistency loss The function processes the first prediction area and the second prediction area to obtain the second loss value.
- the parameter adjustment sub-module includes a weighting processing unit configured to perform weighting processing on the first loss value and the second loss value to obtain the weighted loss value ,
- the parameter adjustment sub-module includes a parameter adjustment unit configured to adjust the network parameters of the original detection model by using the weighted loss value.
- the model can increase the focus on difficult samples, which can help improve the accuracy of the image detection model.
- the first loss value of the collective similarity is obtained, which can make the model fit the pseudo-labeled actual area, which can help improve the accuracy of the image detection model Performance
- the consistency loss function to process the first prediction area and the second prediction area to obtain the second loss value, which can improve the prediction consistency of the original model and the image detection model, which can further improve the performance of the image detection model
- Accuracy By weighting the first loss value and the second loss value, the weighted loss value is obtained, and the weighted loss value is used to adjust the network parameters of the original detection model, which can balance the importance of each loss value in the training process. Thereby, the accuracy of the network parameters can be improved, which in turn can help improve the accuracy of the image detection model.
- the sample medical image further includes the actual region of the labeled organ
- the first detection result further includes the first prediction region of the labeled organ
- the second detection result further includes the second prediction region of the labeled organ.
- the first loss determination submodule is further configured to determine the first loss value of the original detection model by using the difference between the first predicted region and the actual region of the unlabeled organ and the labeled organ
- the second loss determination submodule is also configured In order to use the difference between the first prediction area of the unlabeled organ and the corresponding second prediction area, the second loss value of the original detection model is determined.
- the second detection result also includes the second prediction of the marked organ
- the difference between the first prediction area and the actual area is comprehensively considered, and in the process of determining the second loss value of the original detection model, only the unmarked value is considered
- the difference between the first prediction region of the organ and the corresponding second prediction region can improve the robustness of the consistency constraint of the original detection model and the image detection model, and thus can improve the accuracy of the image detection model.
- the training device 60 of the image detection model further includes a parameter update module configured to update the network parameters of the image detection model by using the network parameters adjusted during this training and several previous trainings.
- the network parameters of the image detection model can be updated by using the network parameters adjusted by the original detection model in this training and several previous trainings, which can further restrict the network parameters in the process of multiple training due to false
- the cumulative error generated by the marked real area improves the accuracy of the image detection model.
- the parameter update module includes a statistics sub-module configured to count the average value of the network parameters adjusted by the original detection model during the current training and several previous trainings, and the parameter update module includes an update sub-module configured to Update the network parameters of the image detection model to the average value of the network parameters of the corresponding original detection model.
- the average value of the network parameters adjusted by the original detection model during the current training and the previous training is counted, and the network parameters of the image detection model are updated to the average value of the network parameters of the corresponding original detection model.
- the image acquisition module 61 includes an image acquisition sub-module configured to acquire a medical image to be pseudo-labeled, wherein at least one unlabeled organ exists in the medical image to be pseudo-labeled, and the image acquisition module 61 includes a single-organ detection sub-module , Is configured to detect pseudo-labeled medical images using a single-organ detection model corresponding to each unlabeled organ to obtain the organ prediction area of each unlabeled organ.
- the image acquisition module 61 includes a pseudo-labeled sub-module and is configured In order to pseudo-label the organ prediction region of the unlabeled organ as the actual region of the unlabeled organ, and use the pseudo-labeled medical image to be pseudo-labeled as the sample medical image.
- the organ prediction area of the non-labeled organ is pseudo-labeled as the actual area of the unlabeled organ, and the pseudo-labeled medical image after the pseudo-labeling is used as the sample medical image.
- the single-organ detection model can be used to eliminate the need for manual pairing. The workload of organ labeling can help reduce the labor cost of training an image detection model for multi-organ detection and improve the efficiency of training.
- the medical image to be pseudo-labeled includes at least one labeled organ
- the image acquisition module 61 further includes a single organ training sub-module configured to use the medical image to be pseudo-labeled to use the labeled organ in the medical image to be pseudo-labeled.
- the corresponding single-organ detection model is trained.
- the medical image to be pseudo-labeled includes at least one labeled organ
- the single-organ detection model corresponding to the labeled organ in the pseudo-labeled medical image is trained by using the medical image to be pseudo-labeled, which can improve the single organ
- the accuracy of the detection model can thus help to improve the accuracy of subsequent pseudo-labeling, which in turn can help improve the accuracy of the subsequent training image detection model.
- the image acquisition sub-module includes a three-dimensional image acquisition unit configured to acquire a three-dimensional medical image
- the image acquisition sub-module includes a pre-processing unit configured to preprocess the three-dimensional medical image
- the image acquisition sub-module includes an image
- the cropping unit is configured to perform cropping processing on the preprocessed three-dimensional medical image to obtain at least one two-dimensional medical image to be pseudo-labeled.
- the pre-processed three-dimensional medical images are cropped to obtain at least one two-dimensional medical image to be pseudo-labeled, which can be beneficial to obtain Medical images that meet model training can help improve the accuracy of subsequent image detection model training.
- the preprocessing unit is further configured to perform at least one of the following: adjust the voxel resolution of the three-dimensional medical image to a preset resolution; use a preset window value to adjust the voxel value of the three-dimensional medical image Normalize to a preset range; add Gaussian noise to at least part of the voxels of the three-dimensional medical image.
- adjusting the voxel resolution of the three-dimensional medical image to a preset resolution can facilitate subsequent model prediction processing; the preset window value is used to normalize the voxel value of the three-dimensional medical image to a preset Within the range, it can help the model to extract accurate features; adding Gaussian noise to at least part of the voxels of the three-dimensional medical image can help achieve data augmentation, increase data diversity, and improve the accuracy of subsequent model training.
- FIG. 7 is a schematic diagram of a framework of an embodiment of an image detection device provided by an embodiment of the present disclosure.
- the image detection device 70 includes an image acquisition module 71 and an image detection module 72.
- the image acquisition module 71 is configured to acquire a medical image to be detected, wherein the medical image to be detected contains multiple organs;
- the image detection module 72 is configured to use image detection
- the model detects the medicine to be detected to obtain predicted regions of multiple organs; wherein, the image detection model is trained by the training device of the image detection model in any of the above-mentioned image detection model training device embodiments.
- the image detection model trained by the training device of the image detection model in the embodiment of the training device for any of the above-mentioned image detection models is used for detection and detection of medical images to be detected, and the predicted regions of multiple organs are obtained. In the process, improve the detection accuracy.
- FIG. 8 is a schematic diagram of a framework of an embodiment of an electronic device provided by an embodiment of the present disclosure.
- the electronic device 80 includes a memory 81 and a processor 82 that are coupled to each other.
- the processor 82 is configured to execute program instructions stored in the memory 81 to implement the steps of any of the foregoing image detection model training method embodiments, or implement any of the foregoing. Steps in an embodiment of an image detection method.
- the electronic device 80 may include but is not limited to: a microcomputer and a server.
- the electronic device 80 may also include mobile devices such as a notebook computer and a tablet computer, which are not limited herein.
- the processor 82 is configured to control itself and the memory 81 to implement the steps of any of the foregoing image detection model training method embodiments, or implement the steps of any of the foregoing image detection method embodiments.
- the processor 82 may also be referred to as a CPU (Central Processing Unit, central processing unit).
- the processor 82 may be an integrated circuit chip with signal processing capability.
- the processor 82 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application specific integrated circuit (ASIC), a field programmable gate array (Field-Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gates or transistor logic devices, discrete hardware components.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor or the like.
- the processor 82 may be jointly implemented by an integrated circuit chip.
- the above solution can improve the accuracy of detection in the process of multi-organ detection.
- FIG. 9 is a schematic framework diagram of an embodiment of a computer-readable storage medium provided by an embodiment of the present disclosure.
- the computer-readable storage medium 90 stores program instructions 901 that can be executed by the processor.
- the program instructions 901 are configured to implement the steps of any of the foregoing image detection model training method embodiments, or implement any of the foregoing image detection method embodiments. A step of.
- the above solution can improve the accuracy of detection in the process of multi-organ detection.
- the training method of the image detection model or the computer program product of the image detection method provided by the embodiments of the present disclosure includes a computer-readable storage medium storing program code, and the instructions included in the program code can be configured to execute the above method embodiments
- the training method of the image detection model or the steps of the image detection method described in the above please refer to the above method embodiment, which will not be repeated here.
- the embodiments of the present disclosure also provide a computer program, which, when executed by a processor, implements any one of the methods in the foregoing embodiments.
- the computer program product can be implemented by hardware, software or a combination thereof.
- the computer program product is embodied as a computer storage medium.
- the computer program product is embodied as a software product, such as a software development kit (SDK) and so on.
- SDK software development kit
- the disclosed method and device may be implemented in other ways.
- the device implementation described above is only illustrative, for example, the division of modules or units is only a logical function division, and there may be other divisions in the actual implementation process, for example, units or components can be combined or It can be integrated into another system, or some features can be ignored or not implemented.
- the displayed or discussed mutual coupling or direct coupling or communication connection may be indirect coupling or communication connection through some interfaces, devices or units, and may be in electrical, mechanical or other forms.
- the units described as separate components may or may not be physically separate, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of this embodiment.
- the functional units in the various embodiments of the present disclosure may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
- the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
- the integrated unit is implemented in the form of a software functional unit and sold or used as an independent product, it can be stored in a computer readable storage medium.
- the technical solution of the present disclosure essentially or the part that contributes to the prior art or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium.
- a computer device which may be a personal computer, a server, or a network device, etc.
- the aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disks or optical disks and other media that can store program codes. .
- a sample medical image is obtained, and the sample medical image pseudo-labels at least one actual region of an unlabeled organ; the original detection model is used to detect the sample medical image to obtain a first detection including the first predicted region of the unlabeled organ Results; use the image detection model to detect the sample medical image to obtain the second detection result including the second prediction area of the unlabeled organ, the network parameters of the image detection model are determined based on the network parameters of the original detection model; use the first prediction The difference between the area and the actual area and the second prediction area respectively, adjust the network parameters of the original detection model. In this way, the detection accuracy can be improved in the process of multi-organ detection.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Medical Informatics (AREA)
- Radiology & Medical Imaging (AREA)
- Quality & Reliability (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Image Analysis (AREA)
- Apparatus For Radiation Diagnosis (AREA)
Abstract
Description
Claims (19)
- 一种图像检测模型的训练方法,包括:An image detection model training method, including:获取样本医学图像,其中,所述样本医学图像伪标注出至少一个未标注器官的实际区域;Acquiring a sample medical image, wherein the sample medical image pseudo-labels at least one actual region of an unlabeled organ;利用原始检测模型对所述样本医学图像进行检测以得到第一检测结果,其中,所述第一检测结果包括所述未标注器官的第一预测区域;以及,Detecting the sample medical image using an original detection model to obtain a first detection result, wherein the first detection result includes the first prediction region of the unlabeled organ; and,利用图像检测模型对所述样本医学图像进行检测以得到第二检测结果,其中,所述第二检测结果包括所述未标注器官的第二预测区域;所述图像检测模型的网络参数是基于所述原始检测模型的网络参数确定的;Use an image detection model to detect the sample medical image to obtain a second detection result, wherein the second detection result includes the second prediction area of the unlabeled organ; the network parameters of the image detection model are based on the The network parameters of the original detection model are determined;利用所述第一预测区域分别与所述实际区域、所述第二预测区域之间的差异,调整所述原始检测模型的网络参数。Using differences between the first prediction area and the actual area and the second prediction area to adjust the network parameters of the original detection model.
- 根据权利要求1所述的训练方法,其中,所述原始检测模型包括第一原始检测模型和第二原始检测模型,所述图像检测模型包括与所述第一原始检测模型对应的第一图像检测模型和与所述第二原始检测模型对应的第二图像检测模型;The training method according to claim 1, wherein the original detection model includes a first original detection model and a second original detection model, and the image detection model includes a first image detection corresponding to the first original detection model. A model and a second image detection model corresponding to the second original detection model;所述利用原始检测模型对所述样本医学图像进行检测以得到第一检测结果,包括:The using the original detection model to detect the sample medical image to obtain the first detection result includes:分别利用所述第一原始检测模型和所述第二原始检测模型执行所述对所述样本医学图像进行检测以得到第一检测结果的步骤;Using the first original detection model and the second original detection model to perform the step of detecting the sample medical image to obtain a first detection result;所述利用图像检测模型对所述样本医学图像进行检测以得到第二检测结果,包括:The using the image detection model to detect the sample medical image to obtain a second detection result includes:分别利用所述第一图像检测模型和第二图像检测模型执行所述对所述样本医学图像进行检测以得到第二检测结果的步骤;Using the first image detection model and the second image detection model to perform the step of detecting the sample medical image to obtain a second detection result;所述利用所述第一预测区域分别与所述实际区域、所述第二预测区域之间的差异,调整所述原始检测模型的网络参数,包括:The adjusting the network parameters of the original detection model by using the differences between the first prediction area and the actual area and the second prediction area respectively includes:利用所述第一原始检测模型的第一预测区域分别与所述实际区域、所述第二图像检测模型的第二预测区域之间的差异,调整所述第一原始检测模型的网络参数;以及,Using the differences between the first prediction area of the first original detection model and the actual area and the second prediction area of the second image detection model to adjust the network parameters of the first original detection model; and ,利用所述第二原始检测模型的第一预测区域分别与所述实际区域、所述第一图像检测模型的第二预测区域之间的差异,调整所述第二原始检测模型的网络参数。The difference between the first prediction area of the second original detection model and the actual area and the second prediction area of the first image detection model is used to adjust the network parameters of the second original detection model.
- 根据权利要求1或2所述的训练方法,其中,所述利用所述第一预测区域分别与所述实际区域、所述第二预测区域之间的差异,调整所述原始检测模型的网络参数包括:The training method according to claim 1 or 2, wherein the difference between the first prediction area and the actual area and the second prediction area is used to adjust the network parameters of the original detection model include:利用所述第一预测区域和所述实际区域之间的差异,确定所述原始检测模型的第一损失值;以及,Using the difference between the first prediction area and the actual area to determine the first loss value of the original detection model; and,利用所述第一预测区域和所述第二预测区域之间的差异,确定所述原始检测模型的第二损失值;Using the difference between the first prediction area and the second prediction area to determine the second loss value of the original detection model;利用所述第一损失值和所述第二损失值,调整所述原始检测模型的网络参数。Using the first loss value and the second loss value to adjust the network parameters of the original detection model.
- 根据权利要求3所述的训练方法,其中,所述利用所述第一预测区域和所述实际区域之间的差异,确定所述原始检测模型的第一损失值包括以下至少之一:The training method according to claim 3, wherein the determining the first loss value of the original detection model by using the difference between the first prediction area and the actual area comprises at least one of the following:利用焦点损失函数对所述第一预测区域和所述实际区域进行处理,得到焦点第一损失值;Processing the first prediction area and the actual area by using a focus loss function to obtain a first focus loss value;利用集合相似度损失函数对所述第一预测区域和所述实际区域进行处理,得到集合相似度第一损失值。The first prediction area and the actual area are processed by using the collective similarity loss function to obtain the first loss value of the collective similarity.
- 根据权利要求3所述的训练方法,其中,所述利用所述第一预测区域和所述第二预测区域之间的差异,确定所述原始检测模型的第二损失值包括:The training method according to claim 3, wherein the determining the second loss value of the original detection model by using the difference between the first prediction area and the second prediction area comprises:利用一致性损失函数对所述第一预测区域和所述第二预测区域进行处理,得到所述第二损失值。The first prediction area and the second prediction area are processed by using a consistency loss function to obtain the second loss value.
- 根据权利要求3所述的训练方法,其中,所述利用所述第一损失值和所述第二损失值,调整所述原始检测模型的网络参数包括:The training method according to claim 3, wherein the adjusting the network parameters of the original detection model by using the first loss value and the second loss value comprises:对所述第一损失值和所述第二损失值进行加权处理,得到加权损失值;Weighting the first loss value and the second loss value to obtain a weighted loss value;利用所述加权损失值,调整所述原始检测模型的网络参数。Using the weighted loss value, adjust the network parameters of the original detection model.
- 根据权利要求3至6任一项所述的训练方法,其中,所述样本医学图像中还包含已标注器官的实际区域,所述第一检测结果还包括所述已标注器官的第一预测区域,所述第二检测结果还包括所述已标注器官的第二预测区域;The training method according to any one of claims 3 to 6, wherein the sample medical image further includes the actual region of the labeled organ, and the first detection result further includes the first prediction region of the labeled organ , The second detection result further includes a second prediction area of the marked organ;所述利用所述第一预测区域和所述实际区域之间的差异,确定所述原始检测模型的第一损失值,包括:The using the difference between the first prediction area and the actual area to determine the first loss value of the original detection model includes:利用所述未标注器官和所述已标注器官的第一预测区域和所述实际区域之间的差异,确定所述原始检测模型的第一损失值;Determine the first loss value of the original detection model by using the difference between the first prediction area and the actual area of the unlabeled organ and the labeled organ;所述利用所述第一预测区域和所述第二预测区域之间的差异,确定所述原始检测模型的第二损失值,包括:The using the difference between the first prediction area and the second prediction area to determine the second loss value of the original detection model includes:利用所述未标注器官的第一预测区域和对应所述第二预测区域之间的差异,确定所述原始检测模型的第二损失值。Using the difference between the first prediction region of the unlabeled organ and the corresponding second prediction region, the second loss value of the original detection model is determined.
- 根据权利要求1至7任一项所述的训练方法,其中,所述利用所述第一预测区域分别与所述实际区域、所述第二预测区域之间的差异,调整所述原始检测模型的网络参数之后,所述方法还包括:The training method according to any one of claims 1 to 7, wherein the difference between the first prediction area and the actual area and the second prediction area is used to adjust the original detection model After the network parameters of, the method further includes:利用本次训练以及之前若干次训练时调整后的网络参数,对所述图像检测模型的网络参数进行更新。The network parameters of the image detection model are updated by using the network parameters adjusted during this training and several previous trainings.
- 根据权利要求8所述的训练方法,其中,所述利用本次训练以及之前若干次训练时调整后的网络参数,对所述图像检测模型的网络参数进行更新,包括:The training method according to claim 8, wherein said using the network parameters adjusted during this training and several previous trainings to update the network parameters of the image detection model comprises:统计所述原始检测模型在本次训练和之前若干次训练所调整的网络参数的平均值;Count the average values of the network parameters adjusted by the original detection model during this training and several previous trainings;将所述图像检测模型的网络参数更新为对应的所述原始检测模型的所述网络参数的平均值。The network parameter of the image detection model is updated to the average value of the network parameter of the corresponding original detection model.
- 根据权利要求1至9任一项所述的训练方法,其中,所述获取样本医学图像包括:The training method according to any one of claims 1 to 9, wherein said obtaining a sample medical image comprises:获取待伪标注医学图像,其中,所述待伪标注医学图像存在至少一个所述未标注器官;Acquiring a medical image to be pseudo-labeled, where at least one unlabeled organ exists in the medical image to be pseudo-labeled;分别利用与每一所述未标注器官对应的单器官检测模型对所述待伪标注医学图像进行检测,以得到每个所述未标注器官的器官预测区域;Detecting the medical image to be pseudo-labeled by using a single-organ detection model corresponding to each of the unlabeled organs to obtain the organ prediction region of each of the unlabeled organs;将所述未标注器官的器官预测区域伪标注为所述未标注器官的实际区域,并将所述伪标注后的待伪标注医学图像作为所述样本医学图像。The organ prediction region of the unlabeled organ is pseudo-labeled as the actual region of the unlabeled organ, and the pseudo-labeled medical image after the pseudo-labeling is used as the sample medical image.
- 根据权利要求10所述的训练方法,其中,所述待伪标注医学图像包括至少一个已标注器官;所述分别利用与每一所述未标注器官对应的单器官检测模型对所述待伪标注医学图像进行检测之前,所述方法还包括:The training method according to claim 10, wherein the medical image to be pseudo-labeled includes at least one labeled organ; and the single-organ detection model corresponding to each of the unlabeled organs is used for each of the unlabeled organs. Before the medical image is detected, the method further includes:利用所述待伪标注医学图像,对所述待伪标注医学图像中的已标注器官对应的单器官检测模型进行训练。Using the medical image to be pseudo-labeled, the single-organ detection model corresponding to the labeled organ in the medical image to be pseudo-labeled is trained.
- 根据权利要求10所述的训练方法,其中,所述获取待伪标注医学图像,包括:The training method according to claim 10, wherein said acquiring the medical image to be pseudo-labeled comprises:获取三维医学图像,并对所述三维医学图像进行预处理;Acquiring a three-dimensional medical image, and preprocessing the three-dimensional medical image;将预处理后的所述三维医学图像进行裁剪处理,得到至少一个二维的待伪标注医学图像。The preprocessed three-dimensional medical image is cropped to obtain at least one two-dimensional medical image to be pseudo-labeled.
- 根据权利要求12所述的训练方法,其中,所述对所述三维医学图像进行预处理包括以下至少之一:The training method according to claim 12, wherein said preprocessing said three-dimensional medical image comprises at least one of the following:将所述三维医学图像的体素分辨率调整至一预设分辨率;Adjusting the voxel resolution of the three-dimensional medical image to a preset resolution;利用一预设窗值将所述三维医学图像的体素值归一化至预设范围内;Using a preset window value to normalize the voxel value of the three-dimensional medical image to a preset range;在所述三维医学图像的至少部分体素中加入高斯噪声。Gaussian noise is added to at least part of the voxels of the three-dimensional medical image.
- 一种图像检测方法,包括:An image detection method, including:获取待检测医学图像,其中,所述待检测医学图像中包含多个器官;Acquiring a medical image to be tested, wherein the medical image to be tested includes a plurality of organs;利用图像检测模型对所述待检测医学进行检测,得到所述多个器官的预测区域;Use an image detection model to detect the medicine to be detected to obtain the predicted regions of the multiple organs;其中,所述图像检测模型是利用权利要求1至13任一项所述的图像检测模型的训练方法训练得到的。Wherein, the image detection model is obtained by training using the training method of the image detection model according to any one of claims 1 to 13.
- 一种图像检测模型的训练装置,包括:A training device for an image detection model includes:图像获取模块,被配置为获取样本医学图像,其中,所述样本医学图像伪标注出至少一个未标注器官的实际区域;An image acquisition module configured to acquire a sample medical image, wherein the sample medical image pseudo-marks the actual area of at least one unmarked organ;第一检测模块,被配置为利用原始检测模型对所述样本医学图像进行检测以得到第一检测结果,其中,所述第一检测结果包括所述未标注器官的第一预测区域;以及,A first detection module configured to detect the sample medical image using an original detection model to obtain a first detection result, wherein the first detection result includes the first prediction region of the unlabeled organ; and,第二检测模块,被配置为利用图像检测模型对所述样本医学图像进行检测以得到第二检测结果,其中,所述第二检测结果包括所述未标注器官的第二预测区域,所述图像检测模型的网络参数是基于所述原始检测模型的网络参数确定的;The second detection module is configured to use an image detection model to detect the sample medical image to obtain a second detection result, wherein the second detection result includes a second prediction area of the unlabeled organ, and the image The network parameters of the detection model are determined based on the network parameters of the original detection model;参数调整模块,被配置为利用所述第一预测区域分别与所述实际区域、所述第二预测区域之间的差异,调整所述原始检测模型的网络参数。The parameter adjustment module is configured to adjust the network parameters of the original detection model by using the differences between the first prediction area and the actual area and the second prediction area, respectively.
- 一种图像检测装置,包括:An image detection device includes:图像获取模块,被配置为获取待检测医学图像,其中,所述待检测医学图像中包含多个器官;An image acquisition module configured to acquire a medical image to be detected, wherein the medical image to be detected contains a plurality of organs;图像检测模块,被配置为利用图像检测模型对所述待检测医学进行检测,得到所述多个器官的预测区域;An image detection module configured to detect the medicine to be detected by using an image detection model to obtain the predicted regions of the multiple organs;其中,所述图像检测模型是利用权利要求15所述的图像检测模型的训练装置训练得到的。Wherein, the image detection model is obtained by training using the image detection model training device of claim 15.
- 一种电子设备,包括相互耦接的存储器和处理器,所述处理器被配置为执行所述存储器中存储的程序指令,以实现权利要求1至13任一项所述的图像检测模型的训练方法,或实现权利要求14所述的图像检测方法。An electronic device comprising a memory and a processor coupled to each other, the processor being configured to execute program instructions stored in the memory to implement the training of the image detection model according to any one of claims 1 to 13 Method, or implement the image detection method of claim 14.
- 一种计算机可读存储介质,其上存储有程序指令,所述程序指令被处理器执行时实现权利要求1至13任一项所述的图像检测模型的训练方法,或实现权利要求14所述的图像检测方法。A computer-readable storage medium, on which program instructions are stored, when the program instructions are executed by a processor, implement the method for training an image detection model according to any one of claims 1 to 13, or implement the method described in claim 14 Image detection method.
- 一种计算机程序,包括计算机可读代码,在所述计算机可读代码在电子设备中运行的过程中,所述电子设备中的处理器执行用于实现权利要求1至13任一项所述的图像检测模型的训练方法,或实现权利要求14所述的图像检测方法。A computer program, comprising computer readable code, in the process of running the computer readable code in an electronic device, the processor in the electronic device executes for realizing any one of claims 1 to 13 The training method of the image detection model, or the realization of the image detection method of claim 14.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020217043241A KR20220016213A (en) | 2020-04-30 | 2020-12-28 | Image detection method and related model training method and related apparatus and apparatus |
JP2021576932A JP2022538137A (en) | 2020-04-30 | 2020-12-28 | Image detection method, related model training method, and related devices and equipment |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010362766.XA CN111539947B (en) | 2020-04-30 | 2020-04-30 | Image detection method, related model training method, related device and equipment |
CN202010362766.X | 2020-04-30 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021218215A1 true WO2021218215A1 (en) | 2021-11-04 |
Family
ID=71967825
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/140325 WO2021218215A1 (en) | 2020-04-30 | 2020-12-28 | Image detection method and relevant model training method, relevant apparatuses, and device |
Country Status (5)
Country | Link |
---|---|
JP (1) | JP2022538137A (en) |
KR (1) | KR20220016213A (en) |
CN (1) | CN111539947B (en) |
TW (1) | TW202145249A (en) |
WO (1) | WO2021218215A1 (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114391828A (en) * | 2022-03-01 | 2022-04-26 | 郑州大学 | Active psychological nursing intervention system for stroke patient |
CN117041531A (en) * | 2023-09-04 | 2023-11-10 | 无锡维凯科技有限公司 | Mobile phone camera focusing detection method and system based on image quality evaluation |
Families Citing this family (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111539947B (en) * | 2020-04-30 | 2024-03-29 | 上海商汤智能科技有限公司 | Image detection method, related model training method, related device and equipment |
CN112132206A (en) * | 2020-09-18 | 2020-12-25 | 青岛商汤科技有限公司 | Image recognition method, training method of related model, related device and equipment |
CN113850179A (en) * | 2020-10-27 | 2021-12-28 | 深圳市商汤科技有限公司 | Image detection method, and training method, device, equipment and medium of related model |
CN112200802B (en) * | 2020-10-30 | 2022-04-26 | 上海商汤智能科技有限公司 | Training method of image detection model, related device, equipment and storage medium |
CN112669293A (en) * | 2020-12-31 | 2021-04-16 | 上海商汤智能科技有限公司 | Image detection method, training method of detection model, related device and equipment |
CN112785573A (en) * | 2021-01-22 | 2021-05-11 | 上海商汤智能科技有限公司 | Image processing method and related device and equipment |
CN112749801A (en) * | 2021-01-22 | 2021-05-04 | 上海商汤智能科技有限公司 | Neural network training and image processing method and device |
CN114049344A (en) * | 2021-11-23 | 2022-02-15 | 上海商汤智能科技有限公司 | Image segmentation method, training method of model thereof, related device and electronic equipment |
CN114429459A (en) * | 2022-01-24 | 2022-05-03 | 上海商汤智能科技有限公司 | Training method of target detection model and corresponding detection method |
CN114155365B (en) * | 2022-02-07 | 2022-06-14 | 北京航空航天大学杭州创新研究院 | Model training method, image processing method and related device |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090116737A1 (en) * | 2007-10-30 | 2009-05-07 | Siemens Corporate Research, Inc. | Machine Learning For Tissue Labeling Segmentation |
CN109166107A (en) * | 2018-04-28 | 2019-01-08 | 北京市商汤科技开发有限公司 | A kind of medical image cutting method and device, electronic equipment and storage medium |
CN109658419A (en) * | 2018-11-15 | 2019-04-19 | 浙江大学 | The dividing method of organella in a kind of medical image |
CN110097557A (en) * | 2019-01-31 | 2019-08-06 | 卫宁健康科技集团股份有限公司 | Automatic medical image segmentation method and system based on 3D-UNet |
CN110188829A (en) * | 2019-05-31 | 2019-08-30 | 北京市商汤科技开发有限公司 | The training method of neural network, the method for target identification and Related product |
CN111539947A (en) * | 2020-04-30 | 2020-08-14 | 上海商汤智能科技有限公司 | Image detection method, training method of related model, related device and equipment |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018033154A1 (en) * | 2016-08-19 | 2018-02-22 | 北京市商汤科技开发有限公司 | Gesture control method, device, and electronic apparatus |
CN108229267B (en) * | 2016-12-29 | 2020-10-16 | 北京市商汤科技开发有限公司 | Object attribute detection, neural network training and region detection method and device |
JP6931579B2 (en) * | 2017-09-20 | 2021-09-08 | 株式会社Screenホールディングス | Live cell detection methods, programs and recording media |
EP3474192A1 (en) * | 2017-10-19 | 2019-04-24 | Koninklijke Philips N.V. | Classifying data |
JP7325414B2 (en) * | 2017-11-20 | 2023-08-14 | コーニンクレッカ フィリップス エヌ ヴェ | Training a First Neural Network Model and a Second Neural Network Model |
JP7066385B2 (en) * | 2017-11-28 | 2022-05-13 | キヤノン株式会社 | Information processing methods, information processing equipment, information processing systems and programs |
CN109086656B (en) * | 2018-06-06 | 2023-04-18 | 平安科技(深圳)有限公司 | Airport foreign matter detection method, device, computer equipment and storage medium |
CN109523526B (en) * | 2018-11-08 | 2021-10-22 | 腾讯科技(深圳)有限公司 | Tissue nodule detection and model training method, device, equipment and system thereof |
CN110148142B (en) * | 2019-05-27 | 2023-04-18 | 腾讯科技(深圳)有限公司 | Training method, device and equipment of image segmentation model and storage medium |
JP2021039748A (en) * | 2019-08-30 | 2021-03-11 | キヤノン株式会社 | Information processor, information processing method, information processing system, and program |
CN111028206A (en) * | 2019-11-21 | 2020-04-17 | 万达信息股份有限公司 | Prostate cancer automatic detection and classification system based on deep learning |
CN111062390A (en) * | 2019-12-18 | 2020-04-24 | 北京推想科技有限公司 | Region-of-interest labeling method, device, equipment and storage medium |
CN110969245B (en) * | 2020-02-28 | 2020-07-24 | 北京深睿博联科技有限责任公司 | Target detection model training method and device for medical image |
-
2020
- 2020-04-30 CN CN202010362766.XA patent/CN111539947B/en active Active
- 2020-12-28 JP JP2021576932A patent/JP2022538137A/en active Pending
- 2020-12-28 KR KR1020217043241A patent/KR20220016213A/en active Search and Examination
- 2020-12-28 WO PCT/CN2020/140325 patent/WO2021218215A1/en active Application Filing
-
2021
- 2021-03-16 TW TW110109420A patent/TW202145249A/en unknown
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20090116737A1 (en) * | 2007-10-30 | 2009-05-07 | Siemens Corporate Research, Inc. | Machine Learning For Tissue Labeling Segmentation |
CN109166107A (en) * | 2018-04-28 | 2019-01-08 | 北京市商汤科技开发有限公司 | A kind of medical image cutting method and device, electronic equipment and storage medium |
CN109658419A (en) * | 2018-11-15 | 2019-04-19 | 浙江大学 | The dividing method of organella in a kind of medical image |
CN110097557A (en) * | 2019-01-31 | 2019-08-06 | 卫宁健康科技集团股份有限公司 | Automatic medical image segmentation method and system based on 3D-UNet |
CN110188829A (en) * | 2019-05-31 | 2019-08-30 | 北京市商汤科技开发有限公司 | The training method of neural network, the method for target identification and Related product |
CN111539947A (en) * | 2020-04-30 | 2020-08-14 | 上海商汤智能科技有限公司 | Image detection method, training method of related model, related device and equipment |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114391828A (en) * | 2022-03-01 | 2022-04-26 | 郑州大学 | Active psychological nursing intervention system for stroke patient |
CN117041531A (en) * | 2023-09-04 | 2023-11-10 | 无锡维凯科技有限公司 | Mobile phone camera focusing detection method and system based on image quality evaluation |
CN117041531B (en) * | 2023-09-04 | 2024-03-15 | 无锡维凯科技有限公司 | Mobile phone camera focusing detection method and system based on image quality evaluation |
Also Published As
Publication number | Publication date |
---|---|
CN111539947A (en) | 2020-08-14 |
TW202145249A (en) | 2021-12-01 |
KR20220016213A (en) | 2022-02-08 |
JP2022538137A (en) | 2022-08-31 |
CN111539947B (en) | 2024-03-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021218215A1 (en) | Image detection method and relevant model training method, relevant apparatuses, and device | |
CN109584254B (en) | Heart left ventricle segmentation method based on deep full convolution neural network | |
Bi et al. | Automatic liver lesion detection using cascaded deep residual networks | |
US11941807B2 (en) | Artificial intelligence-based medical image processing method and medical device, and storage medium | |
WO2021128825A1 (en) | Three-dimensional target detection method, method and device for training three-dimensional target detection model, apparatus, and storage medium | |
CN110363760B (en) | Computer system for recognizing medical images | |
Wang et al. | CheXLocNet: Automatic localization of pneumothorax in chest radiographs using deep convolutional neural networks | |
US9142030B2 (en) | Systems, methods and computer readable storage media storing instructions for automatically segmenting images of a region of interest | |
CN109215014B (en) | Training method, device and equipment of CT image prediction model and storage medium | |
EP3961561A1 (en) | Method for designing a module for image segmentation | |
CN109949280B (en) | Image processing method, image processing apparatus, device storage medium, and growth evaluation system | |
Yang et al. | A deep learning segmentation approach in free‐breathing real‐time cardiac magnetic resonance imaging | |
US20220335600A1 (en) | Method, device, and storage medium for lesion segmentation and recist diameter prediction via click-driven attention and dual-path connection | |
KR102328198B1 (en) | Method and apparatus for measuring volume of organ using artificial neural network | |
CN112767504A (en) | System and method for image reconstruction | |
EP3973508A1 (en) | Sampling latent variables to generate multiple segmentations of an image | |
CN111724371A (en) | Data processing method and device and electronic equipment | |
CN116130090A (en) | Ejection fraction measuring method and device, electronic device, and storage medium | |
US20210110520A1 (en) | Method and system for simulating and constructing original medical images from one modality to other modality | |
CN113284145A (en) | Image processing method and device, computer readable storage medium and electronic device | |
CN115862119B (en) | Attention mechanism-based face age estimation method and device | |
CN114787816A (en) | Data enhancement for machine learning methods | |
CN115496703A (en) | Pneumonia area detection method and system | |
US20240177839A1 (en) | Image annotation systems and methods | |
TWI778670B (en) | Method and system for pneumonia area detection |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20933616 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2021576932 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20217043241 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20933616 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20933616 Country of ref document: EP Kind code of ref document: A1 |