CN116008982A - Radar target identification method based on trans-scale feature aggregation network - Google Patents
Radar target identification method based on trans-scale feature aggregation network Download PDFInfo
- Publication number
- CN116008982A CN116008982A CN202211607715.4A CN202211607715A CN116008982A CN 116008982 A CN116008982 A CN 116008982A CN 202211607715 A CN202211607715 A CN 202211607715A CN 116008982 A CN116008982 A CN 116008982A
- Authority
- CN
- China
- Prior art keywords
- time
- frequency
- feature
- radar
- network
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 41
- 230000002776 aggregation Effects 0.000 title claims abstract description 28
- 238000004220 aggregation Methods 0.000 title claims abstract description 28
- 238000010586 diagram Methods 0.000 claims abstract description 52
- 230000000241 respiratory effect Effects 0.000 claims abstract description 22
- 230000029058 respiratory gaseous exchange Effects 0.000 claims abstract description 17
- 238000012545 processing Methods 0.000 claims abstract description 15
- 230000004931 aggregating effect Effects 0.000 claims abstract description 9
- 230000003068 static effect Effects 0.000 claims abstract description 5
- 238000011176 pooling Methods 0.000 claims description 48
- 238000004458 analytical method Methods 0.000 claims description 22
- 238000001125 extrusion Methods 0.000 claims description 22
- 230000001360 synchronised effect Effects 0.000 claims description 16
- 238000013528 artificial neural network Methods 0.000 claims description 12
- 230000006870 function Effects 0.000 claims description 8
- 230000004913 activation Effects 0.000 claims description 5
- 238000007781 pre-processing Methods 0.000 claims description 5
- 238000006073 displacement reaction Methods 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 4
- 238000009825 accumulation Methods 0.000 claims description 3
- 238000011068 loading method Methods 0.000 claims description 3
- 238000013507 mapping Methods 0.000 claims description 3
- 230000009467 reduction Effects 0.000 claims description 3
- 230000009466 transformation Effects 0.000 claims description 3
- 238000001514 detection method Methods 0.000 abstract description 5
- 230000008569 process Effects 0.000 description 7
- 238000001228 spectrum Methods 0.000 description 5
- 241000282326 Felis catus Species 0.000 description 4
- 241001465754 Metazoa Species 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 4
- 241000282472 Canis lupus familiaris Species 0.000 description 3
- 241000282412 Homo Species 0.000 description 3
- 241000282414 Homo sapiens Species 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 230000006399 behavior Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000001914 filtration Methods 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 238000002156 mixing Methods 0.000 description 2
- 230000010355 oscillation Effects 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 241000282887 Suidae Species 0.000 description 1
- 241000282898 Sus scrofa Species 0.000 description 1
- 230000008901 benefit Effects 0.000 description 1
- 230000002457 bidirectional effect Effects 0.000 description 1
- 239000011449 brick Substances 0.000 description 1
- 230000015556 catabolic process Effects 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 230000000052 comparative effect Effects 0.000 description 1
- 230000000295 complement effect Effects 0.000 description 1
- 238000006731 degradation reaction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000002592 echocardiography Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 230000036039 immunity Effects 0.000 description 1
- 239000002184 metal Substances 0.000 description 1
- 230000035772 mutation Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000000149 penetrating effect Effects 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 230000002441 reversible effect Effects 0.000 description 1
- 238000012552 review Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 229920006395 saturated elastomer Polymers 0.000 description 1
- 230000003595 spectral effect Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 230000008685 targeting Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/10—Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation
Landscapes
- Radar Systems Or Details Thereof (AREA)
Abstract
The invention discloses a radar target identification method based on a trans-scale feature aggregation network, which comprises the following steps: collecting skin vibration echo generated by static target respiration by using a radar to obtain a radar echo signal; processing the radar echo signal by using an SSST algorithm to generate a time-frequency image containing respiratory characteristics; dividing and aggregating the time-frequency images with different scales by using a feature pyramid network, and extracting micro Doppler feature images from feature images with different scales output by the feature pyramid by using a CSFA network; and inputting the aggregated micro Doppler characteristic diagram into a softMax classifier to obtain a static target identification result. According to the invention, the radar echo signals of the stationary targets are processed through the SSST algorithm, so that clutter such as multipath interference and the like can be suppressed, and a high-resolution time-frequency diagram with breathing characteristics is generated; meanwhile, the micro Doppler characteristic map in the time-frequency map is extracted through the trans-scale characteristic aggregation network, so that the identification accuracy of the target is improved, and the efficiency of radar detection of the target is improved.
Description
Technical Field
The invention belongs to the technical field of radar signal processing, and particularly relates to a radar target identification method based on a trans-scale feature aggregation network.
Background
The ultra-wideband continuous wave radar is an emerging sensing technology, and the electromagnetic wave signal emitted by the ultra-wideband continuous wave radar has the characteristics of strong penetrating capacity and high resolution, can reflect the characteristics of scattering point energy distribution and the like of a target, and can be used for detecting vital sign signals.
In order to avoid the degradation of target identification accuracy caused by interference of radar echo signals such as multipath clutter and the like caused by the environment where a stationary target is located, it is important to improve the resolution of a target image. Currently, MSRA (Multiscale Residual Attention, multi-scale residual attention) based networks are adopted in the related art to realize the identification of stationary targets, and the method comprises three parts of radar signal processing, a multi-scale learning architecture and a residual attention learning mechanism. However, the method utilizes a two-dimensional pseudo-color image to identify a static target, has poor clutter interference suppression effect, is difficult to generate a high-resolution image, causes more invalid features to be extracted from the image, and reduces the accuracy of target identification.
Disclosure of Invention
In order to solve the problems in the prior art, the invention provides a radar target identification method based on a cross-scale feature aggregation network. The technical problems to be solved by the invention are realized by the following technical scheme:
the invention provides a radar target identification method based on a trans-scale feature aggregation network, which comprises the following steps:
collecting skin vibration echo generated by static target respiration by using a radar to obtain a radar echo signal;
processing the radar echo signal by using a synchronous extrusion S-transformation time-frequency analysis algorithm to generate a time-frequency image containing respiratory characteristics;
dividing and aggregating the time-frequency images with different scales by using a feature pyramid network, and extracting micro Doppler feature images from feature images with different scales output by the feature pyramid network by using a CSFA network;
and after generating an aggregate micro-Doppler feature map based on the micro-Doppler feature map, inputting the aggregate micro-Doppler feature map to a SoftMax classifier to obtain an identification result of the radar target.
In one embodiment of the present invention, before the step of generating the aggregated micro doppler profile based on the micro doppler profile, the method further comprises:
acquiring a pre-trained Resnet-18 model, and loading a time-frequency diagram data set model file;
after generating an aggregate micro-Doppler feature map based on the micro-Doppler feature map, inputting the aggregate micro-Doppler feature map to a SoftMax classifier to obtain an identification result of the radar target, wherein the method comprises the following steps of:
and after generating an aggregate micro-Doppler feature map based on the micro-Doppler feature map, inputting the aggregate micro-Doppler feature map into a softMax classifier, so that the softMax classifier compares the aggregate micro-Doppler feature map by utilizing the time-frequency map dataset model file to obtain an identification result of the radar target.
In one embodiment of the present invention, the step of processing the radar echo signal using a synchronous extrusion S-transform time-frequency analysis algorithm to generate a time-frequency image including respiration features further includes:
preprocessing the radar echo signal to obtain a radar echo signal containing respiratory characteristics;
and generating a two-dimensional range profile through time accumulation based on the radar echo signals containing respiratory features.
In one embodiment of the present invention, the step of processing the radar echo signal using a synchronous extrusion S-transform time-frequency analysis algorithm to generate a time-frequency image containing respiratory features includes:
and processing the effective channel data in the two-dimensional range profile by using a synchronous extrusion S-transformation time-frequency analysis algorithm, and aggregating all the effective channel data into a time-frequency image.
In one embodiment of the invention, the radar echo signal is processed using a synchronous extrusion S-transform time-frequency analysis algorithm according to the following formula:
wherein f k 、f c Δf c Respectively representing the discrete frequency of the S transformation, the center frequency of the extrusion section and the bandwidth of the extrusion section, b represents the time axis displacement parameter, and Δf k =f k -f k-1 ,Δf c =f c -f c-1 ,ST(f k B) represents S-transformation of radar echo signals, f c (f k B) represents the instantaneous frequency of the radar echo signal.
In one embodiment of the invention, the CSFA network includes a channel attention model and a spatial attention model;
dividing and aggregating the time-frequency images with different scales by using a feature pyramid network, and extracting micro Doppler feature images from feature images with different scales output by the feature pyramid network by using a CSFA network, wherein the method comprises the following steps:
dividing the time-frequency image into different scales by utilizing a characteristic pyramid network;
input feature map F for each scale i (i=2, 3,4, 5) inputting the channel attention model to obtain a one-dimensional channel attention profile M c1 Thereafter, one-dimensional channel attention profile M c1 And input feature map F i Multiplying to obtain channel characteristic F c1 Channel characteristics F c1 Inputting the two-dimensional space attention model to obtain a two-dimensional space attention characteristic diagram M s1 ;
By combining channel characteristics F c1 And two-dimensional spatial attention profile M s1 Multiplication to obtain a spatial feature F s ;
According to the input characteristic diagram F i The spatial feature F s Multiplying to determine feedback variable F FB And apply the feedback variable F FB Inputting the model into the channel attention model to obtain a one-dimensional channel attention characteristic diagram M c2 Thereafter, one-dimensional channel attention profile M c2 And feedback variable F FB Multiplying to obtain channel characteristic F c2 ;
By combining channel characteristics F c2 Inputting the two-dimensional space attention model to obtain a two-dimensional space attention characteristic diagram M s2 ;
By combining channel characteristics F c2 And two-dimensional spatial attention profile M s2 Multiplying to obtain an input feature map F of the current ith layer scale i Is a micro Doppler characteristic diagram
In one embodiment of the invention, the channel attention model comprises: a first maximum pooling layer, a first average pooling layer and a multi-layer fully connected neural network;
mapping each scale feature of the feature pyramid network F i Inputting the channel attention model to obtain channel characteristicsComprises the steps of:
will input a feature map F i Respectively transmitting to a first maximum pooling layer and a first average pooling layer to obtain a first maximum pooling characteristic diagramAnd a first averaged pooling profile +.>
Pooling the first maximum feature mapAnd a first averaged pooling profile +.>The multi-layer fully-connected neural network with the activation function of ReLU is input according to the forward direction, and two output characteristics of the multi-layer fully-connected neural network are added to obtain the one-dimensional channel injectionForce of intention characteristic diagram M c1 One-dimensional channel attention characteristic diagram M c1 With the current input feature map F i Multiplying to obtain channel characteristic F c1 。
In one embodiment of the invention, the spatial attention model comprises: a second max-pooling layer, a second average pooling layer, and a 5 x 5 convolution kernel;
by combining channel characteristics F c1 Inputting the two-dimensional space attention model to obtain a two-dimensional space attention characteristic diagram M s1 Comprises the steps of:
by combining channel characteristics F c1 Respectively inputting to a second maximum pooling layer and a second average pooling layer to obtain a two-dimensional maximum pooling characteristic diagramTwo-dimensional average pooling feature map->Merging the two space attention characteristic images into a matrix, and generating a two-dimensional space attention characteristic image M after the dimension reduction of a convolution layer with a convolution kernel of 5 multiplied by 5 s1 。
In one embodiment of the invention, the feedback variableRepresenting a dot product operation. Compared with the prior art, the invention has the beneficial effects that:
the invention provides a radar target identification method based on a trans-scale feature aggregation network, which processes radar echo signals of targets through a synchronous extrusion S-transformation time-frequency analysis algorithm, can inhibit clutters such as multipath interference and the like, and further generates a high-resolution time-frequency diagram with breathing features; meanwhile, the micro Doppler feature map in the time-frequency map is extracted through the trans-scale feature aggregation network, so that the identification accuracy of the target is improved, and the method can be used for detecting vital sign signals.
The present invention will be described in further detail with reference to the accompanying drawings and examples.
Drawings
FIG. 1 is a flowchart of a method for radar target identification based on a cross-scale feature aggregation network according to an embodiment of the present invention;
FIG. 2 is a time-frequency image acquisition diagram provided by an embodiment of the present invention;
FIG. 3 is a graph showing contrast of different time-frequency images provided by an embodiment of the present invention;
FIG. 4 is a schematic diagram of the overall architecture of a method for radar target recognition based on a cross-scale feature aggregation network according to an embodiment of the present invention;
FIG. 5 is a schematic diagram of a cross-scale feature aggregation network provided by an embodiment of the present invention;
FIG. 6 is a comparison of micro Doppler feature maps of time-frequency images provided by an embodiment of the present invention;
fig. 7a is a schematic diagram of an experimental scenario provided in an embodiment of the present invention;
FIG. 7b is a schematic diagram of another experimental scenario provided by an embodiment of the present invention;
fig. 8 is a comparison chart of convergence curves of a time-frequency analysis algorithm according to an embodiment of the present invention.
FIG. 9 is a graph comparing accuracy rate curves of human body recognition results of different training models according to an embodiment of the present invention.
Detailed Description
The present invention will be described in further detail with reference to specific examples, but embodiments of the present invention are not limited thereto.
Fig. 1 is a flowchart of a method for identifying radar targets based on a cross-scale feature aggregation network according to an embodiment of the present invention. As shown in fig. 1, an embodiment of the present invention provides a radar target recognition method based on a cross-scale feature aggregation network, including:
s1, acquiring skin micro-vibration echo caused by target respiration by using a radar to obtain a radar echo signal;
s2, processing radar echo signals by using a synchronous extrusion S-transformation time-frequency analysis algorithm, and generating a time-frequency image containing breathing characteristics;
s3, dividing and aggregating time-frequency images with different scales by using a feature pyramid network, and extracting micro Doppler feature images from feature images with various scales output by the feature pyramid network by using a CSFA (cross-scale feature aggregation, trans-scale feature aggregation) network;
s4, after the aggregate micro Doppler feature map is generated based on the micro Doppler feature map, the aggregate micro Doppler feature map is input into a softMax classifier, and an identification result of a radar detection target is obtained.
It should be understood that, since the heartbeat vibration of the target is weak, the characteristic error extracted from the heartbeat signal is large, which affects the accuracy of target identification, the present embodiment selects the respiratory signal caused by the fluctuation of the target chest as the target radar echo signal. Specifically, the radar acquires a radar echo signal by collecting skin micro-vibration echo caused by target respiration, and performs preprocessing operations such as local oscillator mixing, hamming window filtering, inverse fast Fourier transform, moving target display, constant false alarm detection and the like on the radar echo signal. Radar echoes accumulate over time to produce a two-dimensional range profile that can show changes in target position over time, but cannot represent doppler characteristic information of the target, and is relatively weak in noise immunity. In the processing process of radar echo signals, an STFT (Short-time Fourier Transform) algorithm is generally adopted to acquire micro Doppler characteristics, however, the resolution problem of the algorithm can cause spectrum leakage, and a blurring effect can occur to a time-frequency image. In view of this, the embodiment of the present invention generates a high-resolution time-frequency image containing respiratory features by using SSST (Synchrosqueezing S-Transform, synchronous extrusion S-Transform) algorithm, where SSST algorithm is a continuous reversible process for identifying and extracting oscillation components such as time-varying frequency and amplitude from a uniformly sampled signal, which can overcome the problem of time and frequency expansion in the conventional time-frequency analysis method, and is beneficial to improving the spectral resolution, so as to extract a micro doppler feature map with higher resolution later. Meanwhile, compared with the time-frequency image generated by STFT, the time-frequency image generated by SSST has obviously improved frequency focusing property, and the time-frequency image of the respiratory feature has clear envelope and no frequency mutation item.
Optionally, using SSST time-frequency analysis algorithm, the radar echo signal is processed according to the following formula:
wherein f k 、f c Δf c Respectively representing the discrete frequency of the S transformation, the center frequency of the extrusion section and the bandwidth of the extrusion section, b represents the time axis displacement parameter, and Δf k =f k -f k-1 ,Δf c =f c -f c-1 ,ST(f k B) represents S-transformation of radar echo signals, f c (f k B) represents the instantaneous frequency of the radar echo signal. The representation transforms the center frequency f on the spectrum c Surrounding frequency interval [ f c -0.5Δf c ,f c +0.5Δf c ]The frequency spectrum in (a) is superimposed and placed at the central frequency f c The resolution is enhanced by compressing the S-transform time spectrum of a frequency bin to a frequency point.
Specifically, the step frequency waveform s (t) of radar emissions may be written as follows:
where N represents the total number of frequency bins of the stepped frequency signal, T represents the duration of each frequency bin, f L Representing the starting carrier frequency, Δf represents the frequency bandwidth, and the function of rect (·) is represented as follows:
the radar return signal r (t) reflected from the target may be expressed as follows:
wherein K represents target respirationThe number of frequency points τ i Indicating the round trip time of the ith scattering point, r ω (t) and r n (t) represents the reflected echo and other environmental disturbances and noise, respectively, of the wall.
First, the S-transform of radar return signal r (t) may be expressed as follows:
where t represents time, f represents frequency, and b represents a time axis displacement parameter. Considering the time-spectral energy distribution of the signal at f=f 0 At, but with the actual time spectrum at f 0 There is a spurious band nearby, so the instantaneous frequency expression of the radar echo signal r (t) is:
finally, the synchronous extrusion S-transform of the radar return signal r (t) can be expressed as follows:
SSST (f) calculated according to the above formula c B) obtaining the corresponding time-frequency image with the breathing characteristic.
The step of processing the radar echo signal by using a synchronous extrusion S-transformation time-frequency analysis algorithm and generating a time-frequency image containing respiratory characteristics is preceded by the following steps:
preprocessing the radar echo signal to obtain a radar echo signal containing respiratory characteristics;
and generating a two-dimensional range profile through time accumulation based on the radar echo signals containing respiratory features.
Fig. 2 is a time-frequency image acquisition diagram provided in an embodiment of the present invention. Specifically, as shown in fig. 2, the present embodiment acquires a radar echo signal including a respiratory feature by performing preprocessing operations such as local oscillator mixing, hamming window filtering, inverse fast fourier transform, moving object display, and constant false alarm detection on the radar echo signal. The radar echo signals containing the breathing characteristics are accumulated in time to generate a two-dimensional range profile which can display the change condition of the position of the target along with time, but cannot represent Doppler characteristic information of the target, and has weak noise resistance. The data of the effective channels are selected from the two-dimensional range profile and processed by using an SSST time-frequency analysis algorithm, so that the corresponding time-frequency image of each channel is obtained, and then the data of all the effective channels are aggregated to generate a final time-frequency image.
Fig. 3 is a comparison chart of different time-frequency images provided by the embodiment of the present invention, where the first behavior is a time-frequency image generated by using an SSST time-frequency analysis algorithm, the second behavior is a time-frequency image generated by using an STFT time-frequency analysis algorithm, the target corresponding to the (a) column is a stationary person, the target corresponding to the (b) column is a stationary pig, the target corresponding to the (c) column is a stationary dog, and the target corresponding to the (d) column is a stationary cat. Obviously, as shown in fig. 3, the time-frequency image generated by SSST is clearer than the image envelope of the time-frequency image generated by STFT, the frequency focusing performance is obviously improved, and meanwhile, the SSST time-frequency analysis algorithm relieves the blurring effect of the time-frequency image, and each oscillation component of the radar echo signal can be well concentrated in the time-frequency image.
Although the targets can be distinguished according to different respiratory features in the time-frequency image, there is a problem of overlapping intervals between respiratory features of each target, resulting in difficulty in identifying different targets. Therefore, in order to make the respiratory features contained in the generated time-frequency image meet the requirement of target recognition, the embodiment of the invention needs to further extract the micro-doppler feature map in the time-frequency image through the neural network.
Fig. 4 is a schematic overall architecture of a radar target recognition method based on a cross-scale feature aggregation network according to an embodiment of the present invention, and fig. 5 is a schematic structural diagram of the cross-scale feature aggregation network according to an embodiment of the present invention. Optionally, referring to fig. 4-5, the csfa network includes a channel attention model and a spatial attention model;
in the step S3, the steps of dividing and aggregating the time-frequency images by using the feature pyramid network to obtain different scales, and extracting the micro doppler feature map from the output feature map of each scale processed by the feature pyramid network by using the CSFA network include:
s301, dividing time-frequency images into different scales by utilizing a characteristic pyramid network;
s302, dividing the input feature map F of each scale i Inputting the channel attention model to obtain a one-dimensional channel attention feature map M c1 Thereafter, one-dimensional channel attention profile M c1 And input feature map F i Multiplying to obtain channel characteristic F c1 Channel characteristics F c1 Input into a spatial attention model to obtain a two-dimensional spatial attention characteristic diagram M s1 ;
S303, characterizing the channel F c1 And two-dimensional spatial attention profile M s1 Multiplication to obtain a spatial feature F s ;
S304, according to the input feature diagram F i Spatial characteristics F s Multiplying to determine feedback variable F FB And will feed back variable F FB Inputting the model into a channel attention model to obtain a one-dimensional channel attention characteristic diagram M c2 Thereafter, one-dimensional channel attention profile M c2 And feedback variable F FB Multiplying to obtain channel characteristic F c2 ;
S305, channel characteristic F c2 Input into a spatial attention model to obtain a two-dimensional spatial attention characteristic diagram M s2 ;
S306, channel characteristic F c2 And two-dimensional spatial attention profile M s2 Multiplying to obtain an input feature map F of the current ith layer scale i Is a micro Doppler characteristic diagram
It should be understood that the breathing characteristics included in the time-frequency image are time-varying, and the problem of overlapping intervals in the time-frequency image is caused by the limitation of the distance resolution, so that the key to solve the problem of object identification is to extract the differential breathing characteristics of different objects from the time-frequency image.
In this embodiment, after obtaining the micro doppler feature map, the input feature map F of the current i-th layer scale is also needed i Is a micro Doppler characteristic diagramMicro Doppler characteristic diagram +.>Aggregation is carried out in an up-sampling mode, and an aggregate micro Doppler characteristic Map is obtained i-1 . Further, the aggregated micro Doppler feature map is input to a softMax classifier, and a radar target identification result is obtained.
The inventor finds that the performance of the depth CNN is better than that of the shallow CNN in the research process, and micro Doppler features can be captured in a layered manner from a time-frequency image so as to obtain good classification performance. However, the depth CNN has a problem that the accuracy is always in a saturated state, which hinders the process of feature optimization. In order to solve the problem, the embodiment uses the depth residual error learning network to extract the characteristic information of the image, and simultaneously, the convolution characteristics of different layers can improve the complementary characteristic information, thereby being beneficial to improving the identification performance of people and animals.
As shown in FIG. 4, resnet-18 has better feature extraction capability on a smaller image dataset and relatively small performance overhead, so in this embodiment, the Resnet-18 model is used as a Backbone to extract convolution features, and the feature pyramid network extracts micro Doppler feature maps from the time-frequency image. Specifically, as shown in fig. 5, the bottom-up flow line is the forward processing of the network, and each extracted feature layer is the output of a layer on the same level. The time-frequency image is based on the input of a feature pyramid network of Resnet-18, the output of which is a 4-layer feature pyramid obtained by applying steps {4,8,16,32}, i.e., P2, P3, P4, P5. And in the flow line from top to bottom, the feature images output by the CSFA network are aggregated by upsampling the high-level semantic feature information. Of these, the high-level feature map contains more information about respiratory features and the low-level feature map contains more information about respiratory feature details.
The feature pyramid network can realize multi-feature representation and improve the image target identification performance. However, the high-level feature map ignores information about the breathing feature details, and the low-level feature map ignores information about the breathing features. To solve this problem, the present embodiment designs a CSFA architecture to extract micro doppler feature patterns in time-frequency images. The network is mainly used for carrying out feature aggregation and feature enhancement on each micro Doppler feature, and simultaneously focusing on the interesting depth feature and suppressing unnecessary features from the channel and the space response, so that a micro Doppler feature map is effectively extracted from a time-frequency image.
Illustratively, the CSFA network includes a channel attention model and a spatial attention model, the channel attention model including: the spatial attention model comprises a first maximum pooling layer, a first average pooling layer and a multi-layer fully connected neural network, wherein the spatial attention model comprises: a second max-pooling layer, a second average pooling layer, and a 5 x 5 convolution kernel;
mapping each scale feature of the feature pyramid F i Inputting the channel attention model to obtain channel characteristic F ic1 Comprises the steps of:
will input a feature map F i Respectively transmitting to a first maximum pooling layer and a first average pooling layer to obtain a first maximum pooling characteristic diagramAnd a first averaged pooling profile +.>
Pooling the first maximum feature mapAnd a first averaged pooling profile +.>The multi-layer fully-connected neural network with the activation function of ReLU is input according to the forward direction, and two output characteristics of the multi-layer fully-connected neural network are added to obtain a one-dimensional channel attention characteristic diagram M c1 One-dimensional channel attention characteristic diagram M c1 With the current input feature map F i Multiplying to obtain channel characteristic F c1 。
Further, the channel characteristics F c1 Inputting the two-dimensional space attention model to obtain a two-dimensional space attention characteristic diagram M s1 Comprises the steps of:
by combining channel characteristics F c1 Respectively inputting to a second maximum pooling layer and a second average pooling layer to obtain a two-dimensional maximum pooling characteristic diagramTwo-dimensional average pooling feature map->Merging the two space attention characteristic images into a matrix, and generating a two-dimensional space attention characteristic image M after the dimension reduction of a convolution layer with a convolution kernel of 5 multiplied by 5 s1 。
Specifically, a feature map is inputWherein C, H, W respectively represent the input feature images F i Channel number, image height, and image width; for the channel attention model, input feature map F i The first maximum pooling feature map +.>And a first averaged pooling profile +.>Respectively inputting the two feature maps into a multi-layer fully-connected neural network with an activation function of ReLU according to a forward input mode, and adding two output features of the multi-layer fully-connected neural network to obtainCx1×1 one-dimensional channel attention profile M c1 Finally, the attention characteristic diagram M of the one-dimensional channel c1 With the current input feature map F i Multiplying to obtain channel characteristic F c1 。
For the spatial attention model, channel characteristics F c1 As input, respectively passing through a second maximum pooling layer and a second average pooling layer to obtain a two-dimensional maximum pooling characteristic diagramTwo-dimensional average pooling feature map->Combining the two feature images into a matrix, and then reducing the dimension of a convolution layer with a convolution kernel of 5×5 to generate a two-dimensional space attention feature image M with H×W×1 s1 Finally M s1 And F is equal to c1 Multiplication to obtain spatial feature F s The course of the process can be written in the form:
wherein,,representing an input profile, < >>Representing multiplication by array elements. The complete flow of the attention model can be written in the form:
wherein,,sigma represents a Sigmoid function for the juxtaposition operator, < ->And-> Representing the weights of an MLP network, a ReLU activation function followed by W 0 R is the decay ratio, f 5×5 A convolution operation with a convolution kernel size of 5 x 5 is represented.
In the present embodiment, spatial feature F s And current feature map F i Multiplication can obtain feedback variable F FB Determining a feedback variable F FB After that, F FB New input feature map as channel attention and spatial attention model, final output F out Is an output characteristic diagram processed by the model. In this way, the feature information of the current layer can use feedback connections to obtain more micro-doppler feature maps, making the new features more diverse. F (F) FB Can be written in the following form:
feedback variable F FB As the characteristics of the bidirectional information flow, the connection of each characteristic is greatly improved in the space dimension and the channel dimension, so that the high-level characteristics and the low-level characteristics are utilized, and the high-quality target identification is realized. Optionally, before the step of inputting the aggregated micro doppler profile to the SoftMax classifier to obtain the identification result of the target, the method further includes:
obtaining a pre-trained Resnet-18 model, and loading a time-frequency diagram data set model file;
in step S4, the step of inputting the aggregated micro doppler profile to a SoftMax classifier to obtain the identification result of the target includes:
and inputting the aggregated micro Doppler feature map into a SoftMax classifier, so that the SoftMax classifier compares the aggregated micro Doppler feature map with a time-frequency map data set model file, and the result after comparison is the identification result of the target.
In the embodiment, a pre-trained Resnet-18 model can be obtained before target identification, and a time-frequency diagram data set model file is loaded, so that the diversity of a data set is improved by adding 5-10 dB of Gaussian white noise into radar echo; further, the aggregated micro Doppler feature map is input to a SoftMax classifier, the SoftMax classifier compares and evaluates in a time-frequency map data set, and finally, the identification result of the target is output, so that the identification of the human and the animal is completed.
Figure 6 is a comparison of micro-doppler profiles of time-frequency images generated by different modeling methods in accordance with an embodiment of the present invention. As shown in fig. 6, when the CSFA network is introduced, more effective micro-doppler features can be obtained from the time-frequency image, so that breathing characteristics of different targets can be fully reflected.
Further, this example performed four sets of comparative experiments targeting humans and three different animals, respectively. Specifically, in this experiment, the radar works in the frequency range of 1.0-2.0 GHz, the duration of each frequency is 100s, the Pulse Repetition Interval (PRI) is 70ms, and a transmitting-receiving array antenna is used for transmitting and receiving signals to the wall.
Fig. 7a and 7b are schematic diagrams of experimental scenarios provided in an embodiment of the present invention, wherein the thickness of the brick wall is about 0.20m. Referring to fig. 7a-7b, humans, dogs, cats and pigs were divided into 4 groups of subjects, each of which was subjected to 10 experiments, in each of which humans and animals were required to remain stationary as subjects. To ensure that the subjects were at rest in each experiment, the time of experimental detection was set to 90s. It should be noted that, since it cannot be guaranteed that the cat is still, during the experimental test, the cat is put into a plastic box in this embodiment, so that interference caused by metal to the radar echo collection is avoided. In addition, the perception range of the through-wall radar is set to be 0-5 m according to the size of an experimental scene.
All models configured NVIDIA GTX 1080GPU and Intel i5-8400, and the optimizer selected SGD mode with a learning rate of 0.01. The main evaluation index adopted in this embodiment is Average Precision (AP), and the definition mode of the accuracy parameter may be written as follows:
the definition mode of the AP can be written as follows:
where TP represents true positive, FP represents false positive, p represents precision, r represents recall, and p is a function of r, which is the sum of the areas under the curve.
Fig. 8 is a comparison chart of convergence curves of a time-frequency analysis algorithm according to an embodiment of the present invention. As shown in fig. 8, the data set of the STFT algorithm converged after 60 iterations, and the data set of the SST algorithm converged after 20 iterations. The comparison result shows that the data set of the SST algorithm has higher convergence speed than the data set of the STFT algorithm. The trained network includes Resnet-18, resnet-18+CBAM, MSRA network and CSFA network architecture, and the precision curves of the four recognition models are shown in FIG. 9. Compared with other networks, the network model of the invention has higher recognition precision, proves the effectiveness of the CSFA network, and the training result is shown in table 1.
TABLE 1
As shown in table 1, the four methods are arranged in the following order for different target recognition performance: resnet-18, resnet-18+CBAM, MSRA networks and methods for target recognition based on through-the-wall radars provided by the invention. Among them, MSRA networks are based on Resnet-18 networks, which add a multi-scale attention mechanism, thus improving the performance of recognizing objects. Experimental results show that the CSFA network improves the precision by 2.57% compared with the CBAM network, and compared with the MSRA network, the classification accuracy of the invention in human, dog, cat and pig is respectively improved by 1.23%, 0.37%, 2.02% and 1.7%. Therefore, the CSFA network can be obtained to effectively extract micro Doppler features from the time-frequency diagram, and accuracy of time-frequency diagram identification of different targets is improved.
According to the above embodiments, the beneficial effects of the invention are as follows:
the invention provides a radar target identification method based on a trans-scale feature aggregation network, which processes radar echo signals of targets through a synchronous extrusion S-transformation time-frequency analysis algorithm, can inhibit clutters such as multipath interference and the like, and further generates a high-resolution time-frequency diagram with breathing features; meanwhile, the micro Doppler characteristic map in the time-frequency map is extracted through the trans-scale characteristic aggregation network, so that the identification accuracy of the target is improved. In the description of the present invention, the terms "first," "second," and the like are used for descriptive purposes only and are not to be construed as indicating or implying relative importance or implicitly indicating the number of technical features indicated. Thus, a feature defining "a first" or "a second" may explicitly or implicitly include one or more such feature. In the description of the present invention, the meaning of "a plurality" is two or more, unless explicitly defined otherwise.
In the description of the present specification, a description referring to terms "one embodiment," "some embodiments," "examples," "specific examples," or "some examples," etc., means that a particular feature, structure, material, or characteristic described in connection with the embodiment or example is included in at least one embodiment or example of the present invention. In this specification, schematic representations of the above terms are not necessarily directed to the same embodiment or example. Furthermore, the particular features, structures, materials, or characteristics described may be combined in any suitable manner in any one or more embodiments or examples. Further, one skilled in the art can engage and combine the different embodiments or examples described in this specification.
Although the present application has been described herein in connection with various embodiments, other variations to the disclosed embodiments can be understood and effected by those skilled in the art in practicing the claimed application, from a review of the figures, the disclosure, and the appended claims. In the claims, the word "comprising" does not exclude other elements or steps, and the "a" or "an" does not exclude a plurality. The mere fact that certain measures are recited in mutually different dependent claims does not indicate that a combination of these measures cannot be used to advantage.
The foregoing is a further detailed description of the invention in connection with the preferred embodiments, and it is not intended that the invention be limited to the specific embodiments described. It will be apparent to those skilled in the art that several simple deductions or substitutions may be made without departing from the spirit of the invention, and these should be considered to be within the scope of the invention.
Claims (9)
1. A radar target identification method based on a trans-scale feature aggregation network is characterized by comprising the following steps:
collecting skin vibration echo generated by static target respiration by using a radar to obtain a radar echo signal;
processing the radar echo signal by using a synchronous extrusion S-transformation time-frequency analysis algorithm to generate a time-frequency image containing respiratory characteristics;
dividing and aggregating the time-frequency images with different scales by using a characteristic pyramid network, and extracting micro Doppler characteristic images from characteristic images of all scales output by the characteristic pyramid network by using a cross-scale characteristic aggregation CSFA network;
and after generating an aggregate micro-Doppler feature map based on the micro-Doppler feature map, inputting the aggregate micro-Doppler feature map to a SoftMax classifier to obtain an identification result of the radar target.
2. The method of claim 1, further comprising, prior to the step of generating an aggregated micro-doppler profile based on the micro-doppler profile:
acquiring a pre-trained Resnet-18 model, and loading a time-frequency diagram data set model file;
after generating an aggregate micro-Doppler feature map based on the micro-Doppler feature map, inputting the aggregate micro-Doppler feature map to a SoftMax classifier to obtain an identification result of the radar target, wherein the method comprises the following steps of:
and after generating an aggregate micro-Doppler feature map based on the micro-Doppler feature map, inputting the aggregate micro-Doppler feature map into a softMax classifier, so that the softMax classifier compares the aggregate micro-Doppler feature map by utilizing the time-frequency map dataset model file to obtain an identification result of the radar target.
3. The method for radar target recognition based on a cross-scale feature aggregation network according to claim 2, wherein the step of processing the radar echo signal using a synchronous extrusion S-transform time-frequency analysis algorithm to generate a time-frequency image containing respiratory features is preceded by the step of:
preprocessing the radar echo signal to obtain a radar echo signal containing respiratory characteristics;
and generating a two-dimensional range profile through time accumulation based on the radar echo signals containing respiratory features.
4. A method of radar target recognition based on a cross-scale feature aggregation network as claimed in claim 3, wherein the step of processing the radar echo signal using a synchronous extrusion S-transform time-frequency analysis algorithm to generate a time-frequency image containing respiratory features comprises:
and processing the effective channel data in the two-dimensional range profile by using a synchronous extrusion S-transformation time-frequency analysis algorithm, and aggregating all the effective channel data into a time-frequency image.
5. The method for radar target recognition based on a cross-scale feature aggregation network according to claim 4, wherein the radar echo signal is processed by using a synchronous extrusion S-transform time-frequency analysis algorithm according to the following formula:
wherein f k 、f c Δf c Respectively representing the discrete frequency of the S transformation, the center frequency of the extrusion section and the bandwidth of the extrusion section, b represents the time axis displacement parameter, and Δf k =f k -f k-1 ,Δf c =f c -f c-1 ,ST(f k B) represents S-transformation of radar echo signals, f c (f k B) represents the instantaneous frequency of the radar echo signal.
6. The method of claim 1, wherein the CSFA network comprises a channel attention model and a spatial attention model;
dividing and aggregating the time-frequency images with different scales by using a feature pyramid network, and extracting micro Doppler feature images from feature images with different scales output by the feature pyramid network by using a CSFA network, wherein the method comprises the following steps:
dividing the time-frequency image into different scales by utilizing a characteristic pyramid network;
input feature map F of each scale obtained by dividing i (i=2, 3,4, 5) inputting the channel attention model to obtain a one-dimensional channel attention profile M c1 Thereafter, one-dimensional channel attention profile M c1 And input feature map F i Multiplying to obtain channel characteristic F c1 Channel characteristics F c1 Inputting the two-dimensional space attention model to obtain a two-dimensional space attention characteristic diagram M s1 ;
By combining channel characteristics F c1 And two-dimensional spatial attention profile M s1 Multiplication to obtain a spatial feature F s ;
According to the input characteristic diagram F i The spatial feature F s Multiplying to determine feedback variable F FB And apply the feedback variable F FB Inputting the model into the channel attention model to obtain a one-dimensional channel attention characteristic diagram M c2 Thereafter, one-dimensional channel attention profile M c2 And feedback variable F FB Multiplying to obtain channel characteristic F c2 ;
By combining channel characteristics F c2 Inputting the two-dimensional space attention model to obtain a two-dimensional space attention characteristic diagram M s2 ;
By combining channel characteristics F c2 And two-dimensional spatial attention profile M s2 Multiplying to obtain an input feature map F of the current ith layer scale i Micro Doppler profile F of (2) iout 。
7. The method of claim 6, wherein the channel attention model comprises: a first maximum pooling layer, a first average pooling layer and a multi-layer fully connected neural network;
mapping each scale feature of the feature pyramid network F i Inputting the channel attention model to obtain channel characteristicsComprises the steps of:
will be transportedGo into feature map F i Respectively transmitting to a first maximum pooling layer and a first average pooling layer to obtain a first maximum pooling characteristic diagramAnd a first averaged pooling profile +.>
Pooling the first maximum feature mapAnd a first averaged pooling profile +.>The multi-layer fully-connected neural network with the activation function of ReLU is input according to the forward direction, and two output characteristics of the multi-layer fully-connected neural network are added to obtain a one-dimensional channel attention characteristic diagram M c1 One-dimensional channel attention characteristic diagram M c1 With the current input feature map F i Multiplying to obtain channel characteristic F c1 。
8. The method of claim 7, wherein the spatial attention model comprises: a second max-pooling layer, a second average pooling layer, and a 5 x 5 convolution kernel;
by combining channel characteristics F c1 Inputting the two-dimensional space attention model to obtain a two-dimensional space attention characteristic diagram M s1 Comprises the steps of:
by combining channel characteristics F c1 Respectively inputting to a second maximum pooling layer and a second average pooling layer to obtain a two-dimensional maximum pooling characteristic diagramTwo-dimensional average pooling feature map->Merging the two space attention characteristic images into a matrix, and generating a two-dimensional space attention characteristic image M after the dimension reduction of a convolution layer with a convolution kernel of 5 multiplied by 5 s1 。
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211607715.4A CN116008982A (en) | 2022-12-14 | 2022-12-14 | Radar target identification method based on trans-scale feature aggregation network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211607715.4A CN116008982A (en) | 2022-12-14 | 2022-12-14 | Radar target identification method based on trans-scale feature aggregation network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN116008982A true CN116008982A (en) | 2023-04-25 |
Family
ID=86036428
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211607715.4A Pending CN116008982A (en) | 2022-12-14 | 2022-12-14 | Radar target identification method based on trans-scale feature aggregation network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN116008982A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117572379A (en) * | 2024-01-17 | 2024-02-20 | 厦门中为科学仪器有限公司 | Radar signal processing method based on CNN-CBAM shrinkage two-class network |
-
2022
- 2022-12-14 CN CN202211607715.4A patent/CN116008982A/en active Pending
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN117572379A (en) * | 2024-01-17 | 2024-02-20 | 厦门中为科学仪器有限公司 | Radar signal processing method based on CNN-CBAM shrinkage two-class network |
CN117572379B (en) * | 2024-01-17 | 2024-04-12 | 厦门中为科学仪器有限公司 | Radar signal processing method based on CNN-CBAM shrinkage two-class network |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Yang et al. | Omnidirectional motion classification with monostatic radar system using micro-Doppler signatures | |
Liu et al. | Deep learning and recognition of radar jamming based on CNN | |
Zhang et al. | u-DeepHand: FMCW radar-based unsupervised hand gesture feature learning using deep convolutional auto-encoder network | |
Pérez et al. | Deep learning radar object detection and classification for urban automotive scenarios | |
CN110348288A (en) | A kind of gesture identification method based on 77GHz MMW RADAR SIGNAL USING | |
CN115063884B (en) | Millimeter wave radar head action recognition method based on multi-domain fusion deep learning | |
Kılıç et al. | Through‐Wall Radar Classification of Human Posture Using Convolutional Neural Networks | |
CN112137620B (en) | Ultra-wideband radar-based human body weak respiration signal detection method | |
Kim et al. | Radar-based human activity recognition combining range–time–Doppler maps and range-distributed-convolutional neural networks | |
CN113837131A (en) | Multi-scale feature fusion gesture recognition method based on FMCW millimeter wave radar | |
He et al. | Deep learning based human activity classification in radar micro-Doppler image | |
Tang et al. | Human activity recognition based on mixed CNN with radar multi-spectrogram | |
CN116008982A (en) | Radar target identification method based on trans-scale feature aggregation network | |
CN113640768A (en) | Low-resolution radar target identification method based on wavelet transformation | |
Yu | Sonar image target detection based on deep learning | |
Zhang et al. | Temporal-range-doppler features interpretation and recognition of hand gestures using MmW FMCW radar sensors | |
CN113126050B (en) | Life detection method based on neural network | |
Qu et al. | Indoor human behavior recognition method based on wavelet scattering network and conditional random field model | |
Rahman et al. | Multi-frequency rf sensor data adaptation for motion recognition with multi-modal deep learning | |
Erdoğan et al. | Object classification on noise-reduced and augmented micro-doppler radar spectrograms | |
Alkasimi et al. | Geolocation tracking for human identification and activity recognition using radar deep transfer learning | |
Peng et al. | Detailed feature representation and analysis of low frequency UWB radar range profile for improving through-wall human activity recognition | |
CN115061094B (en) | Radar target recognition method based on neural network and SVM | |
Ege | Human activity classification with deep learning using FMCW radar | |
CN114428235B (en) | Spatial inching target identification method based on decision level fusion |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |