CN110717471B - B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner - Google Patents

B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner Download PDF

Info

Publication number
CN110717471B
CN110717471B CN201910984884.1A CN201910984884A CN110717471B CN 110717471 B CN110717471 B CN 110717471B CN 201910984884 A CN201910984884 A CN 201910984884A CN 110717471 B CN110717471 B CN 110717471B
Authority
CN
China
Prior art keywords
target
current frame
frame image
value
obtaining
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910984884.1A
Other languages
Chinese (zh)
Other versions
CN110717471A (en
Inventor
杨雪
李俊芳
朱青云
其他发明人请求不公开姓名
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Affiliated Hospital of University of Qingdao
Original Assignee
Affiliated Hospital of University of Qingdao
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Affiliated Hospital of University of Qingdao filed Critical Affiliated Hospital of University of Qingdao
Priority to CN201910984884.1A priority Critical patent/CN110717471B/en
Publication of CN110717471A publication Critical patent/CN110717471A/en
Application granted granted Critical
Publication of CN110717471B publication Critical patent/CN110717471B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • G06T7/0012Biomedical image inspection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/49Segmenting video sequences, i.e. computational techniques such as parsing or cutting the sequence, low-level clustering or determining units such as shots or scenes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10132Ultrasound image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/07Target detection

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computing Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Radiology & Medical Imaging (AREA)
  • Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
  • Medical Informatics (AREA)
  • Quality & Reliability (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Molecular Biology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The embodiment of the invention provides a B-ultrasonic image target detection method and a B-ultrasonic scanner, wherein the B-ultrasonic image target detection method comprises the following steps: acquiring a video of a detection area in a B-ultrasonic image to be scanned by a B-ultrasonic scanner; obtaining a current frame target area and a composite target area based on the video; the composite target region comprises image characteristics of a target to be detected; obtaining the distance between the current frame target area and the composite target area; obtaining a first target vector based on a current frame target area; obtaining a second target vector based on the composite target region; classifying the first target vector based on the trained support vector machine model to obtain a first target value; classifying the second target vector based on the trained support vector machine model to obtain a second target value; obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as a target value; and if the distance is smaller than the target value, acquiring a target image to be detected based on the current frame target area and the composite target area.

Description

B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner
Technical Field
The invention relates to the technical field of computers, in particular to a B-mode ultrasonic image target detection method based on a support vector machine model and a B-mode ultrasonic scanner.
Background
B-ultrasonic is commonly applied in the medical field, and B-ultrasonic technology is used to detect health data of a human body, view the condition of a good baby, the condition of a thyroid gland, a uterus, a breast, a prostate and the like, which all need to obtain B-ultrasonic images (videos) based on scanning of a B-ultrasonic scanner, and obtain required detection data, such as a detected target, based on the images and videos.
Target detection is a relatively mature computer technology, and various methods such as a region-based target detection method and a texture-based target detection method exist, but the detection accuracy of these traditional methods is low. Currently, the target detection has a good effect of performing target detection based on a neural network. However, the target detection method based on the neural network depends on a large number of training samples, and the target detection method based on the neural network does not consider the difference of the training samples, which may result in an insufficient accuracy of the detection result for field application. In the scene that the target needs to be detected with high precision, such as B-ultrasonic image target detection, the target detection result of the target detection method only depending on the neural network or the traditional target detection method is not satisfactory.
Therefore, a B-mode ultrasonic image target detection method for improving the accuracy of target detection is needed. The B-ultrasonic image target detection method can be used for detecting infant images and lesion position images in B-ultrasonic images.
Disclosure of Invention
The embodiment of the invention provides a B-mode ultrasonic image target detection method based on a support vector machine model, and aims to solve the problems in the prior art.
In a first aspect, an embodiment of the present invention provides a method for detecting a B-mode ultrasonic image target based on a support vector machine model, where the method includes:
acquiring a video of a detection area in a B-ultrasonic image to be scanned by a B-ultrasonic scanner;
obtaining a current frame target area and a composite target area based on the video; the composite target region comprises image characteristics of a target to be detected;
obtaining the distance between the current frame target area and the composite target area;
obtaining a first target vector based on a current frame target area;
obtaining a second target vector based on the composite target region;
classifying the first target vector based on the trained support vector machine model to obtain a first target value;
classifying the second target vector based on the trained support vector machine model to obtain a second target value;
obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as a target value;
and if the distance is smaller than the target value, acquiring a target image to be detected based on the current frame target area and the composite target area.
Optionally, the obtaining a composite target region based on the video includes:
obtaining a residual error region based on the previous frame image and the next frame image;
and obtaining the composite target area based on the residual error area and the current frame image.
Optionally, the obtaining a residual region based on the previous frame image and the next frame image includes: obtaining a front pixel point of the previous frame image and a rear pixel point corresponding to the position of the front pixel point in the subsequent frame image;
obtaining first difference value information between a front pixel point and a rear pixel point through absolute difference value sum operation based on a front block and a rear block corresponding to the front pixel point; wherein the previous block is a region in the previous frame image, and the subsequent block is a region in the subsequent frame image; the front block comprises a plurality of front pixel points, the front pixel points correspond to a plurality of first difference information, and the first difference information forms the residual block according to the corresponding relation with the front pixel points.
Optionally, the obtaining the composite target region based on the residual region and the current frame image includes:
fusing the residual error area with the current frame image to obtain a fused current frame image;
preprocessing the fused current frame image to obtain a composite current frame image;
and obtaining the composite target area based on the composite current frame image.
Optionally, the fusing the residual error region with the current frame image to obtain a fused current frame image includes:
if the pixel value of the pixel point (i, j) in the residual region is equal to the pixel value of the pixel point (i, j) in the current frame image, and the pixel value of the pixel point is a first value, assigning the pixel value of the pixel point (i, j) of the fused current frame image to be a second value;
if the pixel value of the pixel point (i, j) in the residual region is equal to the pixel value of the pixel point (i, j) in the current frame image and the pixel value of the pixel point is not a first value, expanding the pixel point (i, j) of the fused current frame image so that the pixel point (i, j) comprises a fusion channel;
and assigning the fusion channel so as to enable the value of the fusion channel to be a second value, wherein the second value is different from the pixel value of the pixel point and the first value.
Optionally, the method further includes: and storing the fused current frame image.
Optionally, the storing the fused current frame image includes:
aiming at each pixel point in the fused current frame image, dividing the fused current frame image into two storage sections, wherein the two storage sections comprise a first storage section and a second storage area; the first storage section is used for storing the value of the fusion channel, and the second storage section is used for storing the position information and the pixel value of the pixel point in the current frame image.
In a second aspect, an embodiment of the present invention provides a B-ultrasonic scanner, which includes a scanning unit and a processing unit; the scanning part is used for scanning the B ultrasonic video of the area to be detected and sending the B ultrasonic video to the processing part; the processing portion is configured to perform any of the methods described above.
Compared with the prior art, the invention achieves the technical effects that:
the embodiment of the invention provides a B-mode ultrasonic image target detection method based on a support vector machine model and a B-mode ultrasonic scanner, wherein the B-mode ultrasonic image target detection method comprises the following steps: acquiring a video of a detection area in a B-ultrasonic image to be scanned by a B-ultrasonic scanner; obtaining a current frame target area and a composite target area based on the video; the composite target region comprises image characteristics of a target to be detected; obtaining the distance between the current frame target area and the composite target area; obtaining a first target vector based on a current frame target area; obtaining a second target vector based on the composite target region; classifying the first target vector based on the trained support vector machine model to obtain a first target value; classifying the second target vector based on the trained support vector machine model to obtain a second target value; obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as a target value; and if the distance is smaller than the target value, acquiring a target image to be detected based on the current frame target area and the composite target area.
Obtaining a first target vector based on a current frame target area and a second target vector based on a composite target area; classifying the first target vector based on a trained support vector machine model to obtain a first target value, wherein the first target value can accurately represent the characteristic information of the target to be detected in the target area of the current frame; classifying the second target vector based on the trained support vector machine model to obtain a second target value, wherein the second target value can accurately represent the characteristic information of the target to be detected in the composite target region; obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as a target value; the target value can accurately represent the difference of the target to be detected in the current frame target area and the composite target area, and if the distance is smaller than the target value, the target image to be detected is obtained based on the current frame target area and the composite target area, so that the target detection accuracy is improved, and the target to be detected precision is improved. The technical problem of low target detection precision in the prior art is solved, and the technical effect of improving the target detection accuracy is achieved.
Drawings
Fig. 1 shows a flowchart of a B-mode ultrasonic image target detection method based on a support vector machine model according to an embodiment of the present invention.
Fig. 2 is a schematic structural diagram of a first convolutional neural network according to an embodiment of the present invention.
Fig. 3 is a schematic structural diagram illustrating a second convolutional neural network according to an embodiment of the present invention.
Fig. 4 is a schematic structural diagram of a B-mode ultrasonic scanner 100 according to an embodiment of the present invention.
Icon: a 100-B ultrasonic scanner; 110-a scanning section 110; 120-a processing section; 130-a display section; 140-storage section.
Detailed Description
As shown in fig. 1, an embodiment of the present invention provides a B-mode ultrasonic image target detection method based on a support vector machine model. The B-mode ultrasonic image target detection method comprises S101-S109 shown in FIG. 1. S101 to S109 are explained below with reference to fig. 1.
S101: and obtaining a video of a detection area in the B-ultrasonic image to be scanned by the B-ultrasonic scanner.
The B-mode ultrasonic video comprises a plurality of frames of images, and the area to be detected is a diseased area or an area to be detected. The image may be referred to as a B-mode ultrasound image.
S102: a current frame target area and a composite target area are obtained based on the video.
Wherein the composite target region includes image features of the target to be detected.
S103: and obtaining the distance between the current frame target area and the composite target area.
S104: a first target vector is obtained based on a current frame target region.
S105: a second target vector is obtained based on the composite target region.
S106: and classifying the first target vector based on the trained support vector machine model to obtain a first target value.
S107: and classifying the second target vector based on the trained support vector machine model to obtain a second target value.
S108: and obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as the target value.
S109: and if the distance is smaller than the target value, acquiring a target image to be detected based on the current frame target area and the composite target area.
By adopting the scheme, the first target vector is obtained based on the current frame target area, and the second target vector is obtained based on the composite target area; classifying the first target vector based on a trained support vector machine model to obtain a first target value, wherein the first target value can accurately represent the characteristic information of the target to be detected in the target area of the current frame; classifying the second target vector based on the trained support vector machine model to obtain a second target value, wherein the second target value can accurately represent the characteristic information of the target to be detected in the composite target region; obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as a target value; the target value can accurately represent the difference of the target to be detected in the current frame target area and the composite target area, and if the distance is smaller than the target value, the target image to be detected is obtained based on the current frame target area and the composite target area, so that the target detection accuracy is improved, and the target to be detected precision is improved. The technical problem of low target detection precision in the prior art is solved, and the technical effect of improving the target detection accuracy is achieved.
Specifically, for S102, the following may be used: and obtaining a residual error area based on a previous frame image and a next frame image of a current frame image in the B-mode ultrasonic video. And fusing the residual error area with the current frame image to obtain a fused current frame image. And carrying out high-pass filtering on the fused current frame image to obtain a high-frequency current frame image. And carrying out low-pass filtering on the fused current frame image to obtain a low-frequency current frame image. And fusing the high-frequency current frame image and the low-frequency frame image to obtain a composite current frame image. And performing target detection on the current frame image based on the first convolution neural network to obtain a current frame target area. And performing target detection on the composite current frame image based on the second convolutional neural network to obtain a composite target area.
By adopting the scheme, the residual error region is obtained based on the previous frame image and the next frame image of the current frame image in the video, the residual error region and the current frame image are fused to obtain the fused current frame image, and the characteristic information of the current frame image is enhanced. The high-pass filtering is carried out on the fused current frame image to obtain a high-frequency current frame image, the high-frequency current frame image reserves the high-frequency characteristic information of the current frame image, the low-pass filtering is carried out on the fused current frame image to obtain a low-frequency current frame image, the low-frequency current frame image reserves the low-frequency characteristic information of the current frame image, the high-frequency current frame image and the low-frequency frame image are fused to obtain a composite current frame image, the characteristic information in the composite current frame image is enhanced, and meanwhile the fidelity of the characteristic information is improved. Performing target detection on the current frame image based on the first convolution neural network to obtain a current frame target area; and performing target detection on the composite current frame image based on the second convolutional neural network to obtain a composite target region, further improving the probability of the target in the target region, and improving the accuracy of target detection. The method comprises the steps of obtaining the distance between a current frame target area and a composite target area, obtaining a target to be detected based on the current frame target area and the composite target area if the distance is smaller than a target value, combining a target detection result of a traditional neural network and a target detection result of a second neural network with high accuracy, improving the accuracy of target detection and improving the precision of the target to be detected.
The method for obtaining the residual error region based on the previous frame image and the next frame image of the current frame image in the B-mode ultrasonic video specifically comprises the following steps: and obtaining a front pixel point of the previous frame of image and a rear pixel point corresponding to the position of the front pixel point in the next frame of image. And obtaining first difference value information between the front pixel point and the rear pixel point through absolute difference value sum operation based on the front block and the rear block corresponding to the front pixel point. The front block is an area in the previous frame of image and the rear block is an area in the next frame of image. The front block comprises a plurality of front pixel points. When the previous block corresponding to the pixel point is a rectangular block with the previous pixel point as the center and the size is set, for example, the current block is a rectangular block of a block of 2 × 2. If the position of the previous pixel point is at the edge of the current frame image, the previous block comprises a plurality of pixel points which are adjacent to the current pixel point and are obtained by taking the current pixel point as the center, and a determined block formed by the pixel points is the current block. For example, if the front pixel is (0, 0), the front block is a block composed of (0, 0), (0, 1), (1, 0), and (1, 1). Aiming at each pixel point in the previous block, obtaining the difference value between the value of each pixel point and the value of the pixel point corresponding to the position of each pixel point in the later block; and carrying out summation operation on the absolute values of the differences to obtain first difference information, wherein a plurality of front pixel points correspond to a plurality of first difference information, and the plurality of first difference information form a residual block according to the corresponding relation with the front pixel points. Each pixel point in each previous block corresponds to a difference, and specifically, the absolute value of each difference is summed. In order to obtain a rear block corresponding to the position of the front block in the previous frame of image, a rear block corresponding to the position of the front block is obtained in the next frame of image, and each pixel point in the current block corresponds to each pixel point in the rear block in a one-to-one position. The position correspondence refers to position one-to-one correspondence, for example, the position of the front pixel point is the same as the position of the rear pixel point, which is specifically embodied that the value of the position of the front pixel point is the same as the value of the position of the rear pixel point, for example, if the value of the position of the front pixel point is (1, 2) and the value of the position of the rear pixel point is (1, 2), the front pixel point corresponds to the position of the rear pixel point. Thus, the sizes of the front and rear blocks are consistent. Specifically, the first difference information is obtained by the following formula (1).
Figure GDA0002468311310000061
Wherein a (i, j) represents the value of the pixel point (i, j) in the front block corresponding to the front pixel point (m, n), b (i, j) represents the value of the pixel point corresponding to the position of the pixel point (i, j) in the rear block, k represents the number of the pixel points of the front block in the horizontal axis direction, and s1(m, n) represents the first difference information. The method comprises the steps of obtaining the absolute value of an obtained difference value by adopting the value of each pixel point in a front block corresponding to a front pixel point, subtracting the value of each pixel point corresponding to each pixel point of the front block in a rear block corresponding to the front pixel point, summing the absolute values corresponding to the pixel points, obtaining first difference value information, and enabling the obtained residual block to accurately represent the difference of a front frame image relative to a rear frame image.
The method for fusing the residual error region with the current frame image to obtain a fused current frame image specifically comprises the following steps:
firstly, the sum of the pixel value of the pixel point (i, j) in the residual region and the pixel value of the pixel point (i, j + k) in the current frame image is obtained, wherein i, j is a positive integer, and k is an integer greater than or equal to 0. Then, if the sum is greater than 255, the pixel value of the pixel point (i, j) of the fused current frame image is a first difference value, and the first difference value is a difference value between 255 and a remainder of a quotient of the sum and 255. For example, the sum of the pixel value of the pixel point (i, j) in the residual region and the pixel value of the pixel point (i, j + k) in the current frame image is Y, Y >255, and the remainder of Y/255 is X, then the first difference value is equal to 255-X. If the sum is not more than 255, the pixel value of the pixel point (i, j) of the fused current frame image is the sum. Namely, the pixel value of the pixel point (i, j) of the fused current frame image is Y. By adopting the scheme, the characteristics that the pixel values of the current frame image are fused with the pixel values of the residual error region and the pixel points of the current frame image are fused, the characteristics of the fused current frame image are enhanced, meanwhile, the fidelity of the fused current frame image is ensured, and further, a foundation is laid for improving the precision of target detection.
The method comprises the following steps of fusing a high-frequency current frame image and a low-frequency frame image to obtain a composite current frame image, and specifically comprises the following steps: obtaining the pixel value of a pixel point (i, j) in the high-frequency current frame image and the average value of the pixel point (i, j) in the low-frequency current frame image; and determining the average value as the pixel value of the pixel point (i, j) of the composite current frame image. The accuracy of the characteristics of the composite current frame image is improved.
As an optional implementation manner, the high-frequency current frame image and the low-frequency frame image are fused to obtain a composite current frame image, specifically: if the pixel value of the pixel point (i, j) in the high-frequency current frame image is equal to the pixel value of the pixel point (i, j) in the low-frequency current frame image, and the pixel value of the pixel point is a first value, the pixel value of the pixel point (i, j) of the composite current frame image is assigned to be a second value; if the pixel value of the pixel point (i, j) in the high-frequency current frame image is the same as that of the pixel point (i, j) in the low-frequency current frame image, and the pixel value of the pixel point is not a first value, the pixel point (i, j) of the composite current frame image is expanded, so that the pixel point (i, j) comprises a fusion channel; and assigning the fusion channel so as to enable the value of the fusion channel to be a second value, wherein the second value is different from the pixel value of the pixel point and the first value. By adopting the scheme, the obtained composite current frame image comprises the pixel information in the high-frequency current frame image and the pixel information in the low-frequency current frame image, the characteristics of the composite current frame image are enhanced, and the accuracy of the detected target is improved.
In the embodiment of the present invention, the step of performing target detection on the current frame image based on the first convolution neural network to obtain the current frame target area specifically includes: after performing convolution processing on the current frame image for at least one time, obtaining first output data; performing pooling processing on the first output data twice to obtain second output data; after carrying out convolution processing on the second output data for at least two times, obtaining third output data; performing pooling processing on the third output data for at least three times to obtain fourth output data; and classifying the fourth output data to obtain the current target area.
In order to clearly illustrate the step of performing target detection on the current frame image based on the first convolutional neural network to obtain the target region of the current frame, as shown in fig. 2, fig. 2 shows a schematic structural diagram of the first convolutional neural network, that is, the first convolutional neural network includes at least four convolutional layers, at least five pooling layers, and one classification layer.
The first convolutional neural network may have a structure of a convolutional layer, a pooling layer, a convolutional layer, a pooling layer, and a classification layer.
The first convolutional neural network may also have a structure of a convolutional layer, a pooling layer, a convolutional layer, a pooling layer, or a classification layer. Wherein the classification layer may be a support vector machine model. And the pooling layer connected with the classification layer outputs an image characteristic vector, and the classification layer classifies and identifies the image characteristic vector to obtain a current frame target area.
By adopting the scheme, the long convolution kernel pooling operation is carried out on the current frame image, and finally the classification layer carries out classification and identification on the image feature vectors to obtain the current frame target area, so that the accuracy of identifying and detecting the current frame target area is improved.
Optionally, the target detection is performed on the composite current frame image based on the second convolutional neural network to obtain a composite target region, specifically: performing pooling processing on the composite current frame image for at least three times to obtain fifth output data; performing convolution processing on the fifth output data for at least two times to obtain sixth output data; performing pooling processing and convolution processing on the sixth output data to obtain seventh output data; fusing the sixth output data and the seventh output data to obtain eighth output data; performing pooling processing and convolution processing at least twice on the eighth output data to obtain ninth output data; and classifying the ninth output data to obtain a composite target area.
The second convolutional neural network at least comprises six pooling layers, at least five convolutional layers, a fusion layer and a classification layer. A schematic diagram of a second convolutional neural network shown in fig. 3.
The structure of the second convolutional neural network may be a pooling layer, a convolutional layer, a pooling layer, a convolutional layer, a fusion layer, a pooling layer, a convolutional layer, a classification layer. The structure of the second convolutional neural network may be a pooling layer, a convolutional layer, a pooling layer, a convolutional layer, a fusion layer, a pooling layer, a convolutional layer, a sorting layer. Wherein the classification layer may be a support vector machine model as described above. The fusion layer is used for respectively carrying out merging processing on the output of the neurons of the convolution layer connected with the fusion layer, the merging mode can be weighted accumulation or OR operation and the like to obtain a merging processing result, and then the merging processing result is output to the pooling layer for pooling processing. And finally, outputting the region vector by the convolution layer, and classifying the region vector through a classification layer to obtain a composite target region.
By adopting the scheme, the composite current frame image is subjected to multi-length convolution kernel pooling operation, the final convolution layer outputs the region vector, the region vector is classified through the classification layer to obtain the target region, and the precision of identifying and detecting the composite target region is improved.
Optionally, a specific implementation manner of obtaining the distance between the current frame target region and the composite target region is as follows:
acquiring circumscribed circles of a current frame target area and a composite target area; obtaining the Euclidean distance between the circle center of the circumscribed circle of the current frame target area and the circle center of the circumscribed circle of the composite target area; and determining the Euclidean distance as the distance between the current frame target area and the composite target area. The method for obtaining the distance between the current frame target area and the composite target area specifically comprises the following steps: respectively obtaining the gravity centers of a current frame target area and a composite target area; obtaining the Euclidean distance between the gravity center of the current frame target area and the gravity center of the composite target area; and determining the Euclidean distance as the distance between the current frame target area and the composite target area.
The specific implementation of obtaining the target to be detected based on the current frame target region and the composite target region may be as follows: acquiring a cross region of a current frame target region and a composite target region; and determining the cross region as a target to be detected. Thus, the accuracy of the target to be detected is improved.
In order to further improve the accuracy of the target to be detected, before determining the intersection region as the target to be detected, the method for detecting the B-mode ultrasonic image target further includes: and rendering the intersection area. The rendering method may be color rendering according to pathological features and color features of the examined region, for example, if the image is a B-mode ultrasound image of a lung, the rendered object to be detected is a color image of the lung. If the object is a B-mode ultrasound image of the brain, the rendered object to be detected is a gray-scale brain image or a head skeleton image and the like.
In order to improve the rendering accuracy and further improve the target detection accuracy, before the rendering is performed on the intersection region, the B-mode ultrasonic image target detection method further includes: and performing expansion processing on the intersection area so as to enable the size of the intersection area to be close to that of the target to be measured. And expanding the intersection area to enable the size of the intersection area to be close to that of the target to be detected, so that the definition of the rendered intersection area is improved, and the accuracy of target detection is improved.
In the embodiment of the present invention, the target value may be a preset value according to experience. In order to further improve the accuracy of target detection, the target value is obtained by: obtaining a first target vector based on a current frame target area; obtaining a second target vector based on the composite target region; classifying the first target vector based on the trained support vector machine model to obtain a first target value; classifying the second target vector based on the trained support vector machine model to obtain a second target value; and obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as a target value.
As an alternative embodiment, the B-mode ultrasonic image target detection method includes: and obtaining the B ultrasonic video of the area to be detected scanned by the B ultrasonic scanner. And obtaining a residual error area based on a previous frame image and a next frame image of a current frame image in the B-mode ultrasonic video. And fusing the residual error area with the current frame image to obtain a fused current frame image. And carrying out high-pass filtering on the fused current frame image to obtain a high-frequency current frame image. And carrying out low-pass filtering on the fused current frame image to obtain a low-frequency current frame image. And fusing the high-frequency current frame image and the low-frequency frame image to obtain a composite current frame image. And performing target detection on the current frame image based on the first convolution neural network to obtain a current frame target area. A first target vector is obtained based on a current frame target region. And performing target detection on the composite current frame image based on the second convolutional neural network to obtain a composite target area. A second target vector is obtained based on the composite target region. Classifying the first target vector based on the trained support vector machine model to obtain a first target value; classifying the second target vector based on the trained support vector machine model to obtain a second target value; and obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as the target value. And obtaining the distance between the current frame target area and the composite target area. And if the distance is smaller than the target value, obtaining a target image to be detected based on the current frame target area and the composite target area.
As an optional implementation, when the composite target region is stored, each pixel point in the composite target region is divided into two storage sections, where the two storage sections include a first storage section and a second storage region, the first storage section is used to store a value of the fusion channel, and the second storage section is used to store position information and RGB information of the pixel point. Therefore, the storage accuracy is improved, and the fidelity of the stored composite current frame image is ensured. And storing. And aiming at each pixel point, dividing the pixel point into two storage intervals, wherein the first storage interval is used for storing the value of the fusion channel, and the second storage interval is used for storing the position information and the RGB information of the pixel point. Therefore, the storage accuracy is improved, and the fidelity of the stored composite current frame image is ensured.
By adopting the scheme, the residual error region is obtained by adopting the scheme based on the previous frame image and the next frame image of the current frame image in the video, the residual error region and the current frame image are fused to obtain the fused current frame image, and the characteristic information of the current frame image is enhanced. The high-pass filtering is carried out on the fused current frame image to obtain a high-frequency current frame image, the high-frequency current frame image reserves the high-frequency characteristic information of the current frame image, the low-pass filtering is carried out on the fused current frame image to obtain a low-frequency current frame image, the low-frequency current frame image reserves the low-frequency characteristic information of the current frame image, the high-frequency current frame image and the low-frequency frame image are fused to obtain a composite current frame image, the characteristic information in the composite current frame image is enhanced, and meanwhile the fidelity of the characteristic information is improved. Performing target detection on the current frame image based on the first convolution neural network to obtain a current frame target area; and performing target detection on the composite current frame image based on the second convolutional neural network to obtain a composite target region, further improving the probability of the target in the target region, and improving the accuracy of target detection. Obtaining a first target vector based on a current frame target area, obtaining a second target vector based on a composite target area, and classifying the first target vector based on a trained support vector machine model to obtain a first target value; classifying the second target vector based on the trained support vector machine model to obtain a second target value; and obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as the target value. The method comprises the steps of obtaining the distance between a current frame target area and a composite target area, obtaining a target to be detected based on the current frame target area and the composite target area if the distance is smaller than a target value, combining a target detection result of a traditional neural network and a target detection result of a second neural network with high accuracy, improving the accuracy of target detection and improving the precision of the target to be detected. Meanwhile, the target value is obtained based on the target area of the previous frame and the composite target area, and the target value is used as a judgment reference, so that the target detection accuracy can be improved.
The B-ultrasonic image target detection method can be used for detecting the infant image and the lesion position image in the B-ultrasonic image.
On the basis of the above B-mode ultrasonic image target detection method, the embodiment of the present invention further provides a B-mode ultrasonic scanner 100. The B-ultrasonic scanner 100 includes a scanning unit 110, a processing unit 120, and a display unit 130. As shown in fig. 4, the scanning unit 110 is configured to scan a B-mode video of a region to be detected and send the B-mode video to the processing unit 120.
The processing unit 120 is configured to execute any one of the above methods for detecting a target based on a B-mode ultrasound image of a support vector machine model, so as to perform target detection on the B-mode ultrasound image in a B-mode ultrasound video, obtain a target image to be detected, and send the obtained target image to be detected to the display unit 130.
The display unit 130 is used to display an object image to be detected.
When the processing unit 120 executes any of the above methods for detecting a B-mode ultrasonic image based on the support vector machine model, it executes the following functions: and obtaining the B ultrasonic video of the area to be detected scanned by the B ultrasonic scanner. And obtaining a residual error area based on a previous frame image and a next frame image of a current frame image in the B-mode ultrasonic video. And fusing the residual error area with the current frame image to obtain a fused current frame image. And carrying out high-pass filtering on the fused current frame image to obtain a high-frequency current frame image. And carrying out low-pass filtering on the fused current frame image to obtain a low-frequency current frame image. And fusing the high-frequency current frame image and the low-frequency frame image to obtain a composite current frame image. And performing target detection on the current frame image based on the first convolution neural network to obtain a current frame target area. And performing target detection on the composite current frame image based on the second convolutional neural network to obtain a composite target area. And obtaining the distance between the current frame target area and the composite target area. And if the distance is smaller than the target value, acquiring a target image to be detected based on the current frame target area and the composite target area.
By adopting the scheme, the residual error region is obtained by adopting the scheme based on the previous frame image and the next frame image of the current frame image in the video, the residual error region and the current frame image are fused to obtain the fused current frame image, and the characteristic information of the current frame image is enhanced. The high-pass filtering is carried out on the fused current frame image to obtain a high-frequency current frame image, the high-frequency current frame image reserves the high-frequency characteristic information of the current frame image, the low-pass filtering is carried out on the fused current frame image to obtain a low-frequency current frame image, the low-frequency current frame image reserves the low-frequency characteristic information of the current frame image, the high-frequency current frame image and the low-frequency frame image are fused to obtain a composite current frame image, the characteristic information in the composite current frame image is enhanced, and meanwhile the fidelity of the characteristic information is improved.
Performing target detection on the current frame image based on the first convolution neural network to obtain a current frame target area; and performing target detection on the composite current frame image based on the second convolutional neural network to obtain a composite target region, further improving the probability of the target in the target region, and improving the accuracy of target detection.
Obtaining a first target vector based on a current frame target area, obtaining a second target vector based on a composite target area, and classifying the first target vector based on a trained support vector machine model to obtain a first target value; classifying the second target vector based on the trained support vector machine model to obtain a second target value; and obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as the target value. The method comprises the steps of obtaining the distance between a current frame target area and a composite target area, obtaining a target to be detected based on the current frame target area and the composite target area if the distance is smaller than a target value, combining a target detection result of a traditional neural network and a target detection result of a second neural network with high accuracy, improving the accuracy of target detection and improving the precision of the target to be detected. Meanwhile, the target value is obtained based on the target area of the previous frame and the composite target area, and the target value is used as a judgment reference, so that the target detection accuracy can be improved.
Optionally, the B-mode ultrasound scanner 100 further includes a storage unit 140, where the storage unit is configured to store the detected target image and any data in the B-mode ultrasound image target detection method, and specifically, when the composite current frame image is stored, two storage sections are divided for each pixel, where a first storage section is configured to store a value of the fusion channel, and a second storage section is configured to store position information and RGB information of the pixel. Therefore, the accuracy and the effectiveness of storage are improved, and the fidelity of the stored composite current frame image is ensured.

Claims (8)

1. A B-mode ultrasonic image target detection method based on a support vector machine model is characterized by comprising the following steps:
acquiring a video of a detection area in a B-ultrasonic image to be scanned by a B-ultrasonic scanner;
obtaining a current frame target area and a composite target area based on the video; the composite target region comprises image characteristics of a target to be detected;
obtaining the distance between the current frame target area and the composite target area;
obtaining a first target vector based on a current frame target area;
obtaining a second target vector based on the composite target region;
classifying the first target vector based on the trained support vector machine model to obtain a first target value;
classifying the second target vector based on the trained support vector machine model to obtain a second target value;
obtaining the difference between the first target value and the second target value to obtain a target difference value, and taking the target difference value as a target value;
if the distance is smaller than the target value, acquiring a target image to be detected based on the current frame target area and the composite target area; the specific way of obtaining the target image to be detected based on the current frame target area and the composite target area is as follows: acquiring a cross region of a current frame target region and a composite target region; and determining the cross region as a target to be detected.
2. The method of claim 1, wherein obtaining a composite target region based on the video comprises:
obtaining a residual error area based on the previous frame image and the next frame image;
and obtaining the composite target area based on the residual error area and the current frame image.
3. The method of claim 2, wherein obtaining a residual region based on the previous frame image and the next frame image comprises:
obtaining a front pixel point of the previous frame image and a rear pixel point corresponding to the position of the front pixel point in the subsequent frame image;
obtaining first difference value information between a front pixel point and a rear pixel point through absolute difference value sum operation based on a front block and a rear block corresponding to the front pixel point; wherein the previous block is a region in the previous frame image, and the subsequent block is a region in the subsequent frame image; the front block comprises a plurality of front pixel points, the front pixel points correspond to a plurality of first difference information, and the first difference information forms the residual block according to the corresponding relation with the front pixel points.
4. The method of claim 3, wherein obtaining the composite target region based on the residual region and the current frame image comprises:
fusing the residual error area with the current frame image to obtain a fused current frame image;
preprocessing the fused current frame image to obtain a composite current frame image;
and obtaining the composite target area based on the composite current frame image.
5. The method according to claim 4, wherein said fusing the residual region with the current frame image to obtain a fused current frame image comprises:
if the pixel value of the pixel point (i, j) in the residual region is equal to the pixel value of the pixel point (i, j) in the current frame image, and the pixel value of the pixel point is a first value, assigning the pixel value of the pixel point (i, j) of the fused current frame image to be a second value;
if the pixel value of the pixel point (i, j) in the residual region is equal to the pixel value of the pixel point (i, j) in the current frame image and the pixel value of the pixel point is not a first value, expanding the pixel point (i, j) of the fused current frame image so that the pixel point (i, j) comprises a fusion channel;
and assigning the fusion channel so as to enable the value of the fusion channel to be a second value, wherein the second value is different from the pixel value of the pixel point and the first value.
6. The method of claim 5, further comprising:
and storing the fused current frame image.
7. The method of claim 6, wherein storing the fused current frame image comprises:
aiming at each pixel point in the fused current frame image, dividing the fused current frame image into two storage sections, wherein the two storage sections comprise a first storage section and a second storage area; the first storage section is used for storing the value of the fusion channel, and the second storage section is used for storing the position information and the pixel value of the pixel point in the current frame image.
8. A B-ultrasonic scanner is characterized in that the B-ultrasonic scanner comprises a scanning part and a processing part;
the scanning part is used for scanning the B ultrasonic video of the area to be detected and sending the B ultrasonic video to the processing part;
the processing portion is configured to perform the method of any one of claims 1-7.
CN201910984884.1A 2019-10-16 2019-10-16 B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner Active CN110717471B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910984884.1A CN110717471B (en) 2019-10-16 2019-10-16 B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910984884.1A CN110717471B (en) 2019-10-16 2019-10-16 B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner

Publications (2)

Publication Number Publication Date
CN110717471A CN110717471A (en) 2020-01-21
CN110717471B true CN110717471B (en) 2020-07-24

Family

ID=69211767

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910984884.1A Active CN110717471B (en) 2019-10-16 2019-10-16 B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner

Country Status (1)

Country Link
CN (1) CN110717471B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112132480A (en) * 2020-09-29 2020-12-25 姜锡忠 Master and resource matching method and system for big data online education platform

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN100495438C (en) * 2007-02-09 2009-06-03 南京大学 Method for detecting and identifying moving target based on video monitoring
CN101261681B (en) * 2008-03-31 2011-07-20 北京中星微电子有限公司 Road image extraction method and device in intelligent video monitoring
US20120060110A1 (en) * 2010-09-08 2012-03-08 Microsoft Corporation Notification bar user interface control
CN107507221A (en) * 2017-07-28 2017-12-22 天津大学 With reference to frame difference method and the moving object detection and tracking method of mixed Gauss model

Also Published As

Publication number Publication date
CN110717471A (en) 2020-01-21

Similar Documents

Publication Publication Date Title
CN109685060B (en) Image processing method and device
CN111445478B (en) Automatic intracranial aneurysm region detection system and detection method for CTA image
JP3735344B2 (en) Calibration apparatus, calibration method, and calibration program
CN109472820B (en) Monocular RGB-D camera real-time face reconstruction method and device
CN110555908B (en) Three-dimensional reconstruction method based on indoor moving target background restoration
CN106529573A (en) Real-time object detection method based on combination of three-dimensional point cloud segmentation and local feature matching
CN112801074B (en) Depth map estimation method based on traffic camera
JP2014096152A (en) Backfilling points in point cloud
CN112862824A (en) Novel coronavirus pneumonia focus detection method, system, device and storage medium
JP2006524394A (en) Delineation of human contours in images
JPH0935061A (en) Image processing method
CN104217459B (en) A kind of spheroid character extracting method
CN112085675B (en) Depth image denoising method, foreground segmentation method and human motion monitoring method
JP2001067463A (en) Device and method for generating facial picture from new viewpoint based on plural facial pictures different in viewpoint, its application device and recording medium
EP4174792A1 (en) Method for scene understanding and semantic analysis of objects
CN110717910B (en) CT image target detection method based on convolutional neural network and CT scanner
CN111179173B (en) Image splicing method based on discrete wavelet transform and gradient fusion algorithm
JP2001236522A (en) Image processor
EP1374181A1 (en) Image processing method for fitness estimation of a 3d mesh model mapped onto a 3d surface of an object
CN110717471B (en) B-ultrasonic image target detection method based on support vector machine model and B-ultrasonic scanner
CN111738061A (en) Binocular vision stereo matching method based on regional feature extraction and storage medium
CN116052120A (en) Excavator night object detection method based on image enhancement and multi-sensor fusion
Dong et al. Learning stratified 3D reconstruction
CN113729747B (en) Spherical metal marked cone beam CT metal artifact removal system and removal method
JPH05314244A (en) Three-dimensional information extracting method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information
CB03 Change of inventor or designer information

Inventor after: Yang Xue

Inventor after: Li Junfang

Inventor after: Zhu Qingyun

Inventor after: Other inventor requests not to publish the name

Inventor before: Request for anonymity

TA01 Transfer of patent application right
TA01 Transfer of patent application right

Effective date of registration: 20200628

Address after: 266000 Shandong city of Qingdao province Jiangsu City Road No. 16

Applicant after: THE AFFILIATED HOSPITAL OF QINGDAO University

Address before: 610000 15 / F, block a, building 1, Shangding international, Wuhou District, Chengdu, Sichuan Province

Applicant before: Jiang Tongyuan

GR01 Patent grant
GR01 Patent grant