AU2022200439B2 - Multi-modal speech separation method and system - Google Patents
Multi-modal speech separation method and system Download PDFInfo
- Publication number
- AU2022200439B2 AU2022200439B2 AU2022200439A AU2022200439A AU2022200439B2 AU 2022200439 B2 AU2022200439 B2 AU 2022200439B2 AU 2022200439 A AU2022200439 A AU 2022200439A AU 2022200439 A AU2022200439 A AU 2022200439A AU 2022200439 B2 AU2022200439 B2 AU 2022200439B2
- Authority
- AU
- Australia
- Prior art keywords
- speech
- speakers
- spectrogram
- audio
- feature
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000926 separation method Methods 0.000 title claims abstract description 156
- 230000000007 visual effect Effects 0.000 claims abstract description 68
- 238000000034 method Methods 0.000 claims abstract description 46
- 238000000605 extraction Methods 0.000 claims abstract description 45
- 230000004927 fusion Effects 0.000 claims abstract description 38
- 238000012549 training Methods 0.000 claims abstract description 31
- 238000012545 processing Methods 0.000 claims abstract description 14
- 230000001815 facial effect Effects 0.000 claims abstract description 10
- 238000001514 detection method Methods 0.000 claims abstract description 6
- 238000013527 convolutional neural network Methods 0.000 claims description 26
- 230000008569 process Effects 0.000 claims description 16
- 238000007781 pre-processing Methods 0.000 claims description 9
- 230000002123 temporal effect Effects 0.000 claims description 7
- 230000010339 dilation Effects 0.000 claims description 5
- 230000015654 memory Effects 0.000 claims description 5
- 238000004590 computer program Methods 0.000 claims description 4
- 238000005070 sampling Methods 0.000 claims description 3
- 230000005236 sound signal Effects 0.000 claims description 3
- 230000001131 transforming effect Effects 0.000 claims description 3
- 238000007499 fusion processing Methods 0.000 claims description 2
- 230000011218 segmentation Effects 0.000 abstract description 5
- 238000002474 experimental method Methods 0.000 description 13
- 210000003477 cochlea Anatomy 0.000 description 10
- 238000004422 calculation algorithm Methods 0.000 description 9
- 230000000694 effects Effects 0.000 description 8
- 238000005516 engineering process Methods 0.000 description 8
- 238000013135 deep learning Methods 0.000 description 6
- 238000010586 diagram Methods 0.000 description 4
- 239000000284 extract Substances 0.000 description 4
- 230000006872 improvement Effects 0.000 description 4
- 238000012880 independent component analysis Methods 0.000 description 4
- 238000013528 artificial neural network Methods 0.000 description 3
- 210000002469 basement membrane Anatomy 0.000 description 3
- 238000011161 development Methods 0.000 description 3
- 230000018109 developmental process Effects 0.000 description 3
- 238000011156 evaluation Methods 0.000 description 3
- 238000012986 modification Methods 0.000 description 3
- 230000004048 modification Effects 0.000 description 3
- 230000000306 recurrent effect Effects 0.000 description 3
- 230000006403 short-term memory Effects 0.000 description 3
- 238000004458 analytical method Methods 0.000 description 2
- 210000000721 basilar membrane Anatomy 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000000873 masking effect Effects 0.000 description 2
- 238000010606 normalization Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000009466 transformation Effects 0.000 description 2
- 238000000844 transformation Methods 0.000 description 2
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 1
- 241000282412 Homo Species 0.000 description 1
- 238000007476 Maximum Likelihood Methods 0.000 description 1
- 235000000434 Melocanna baccifera Nutrition 0.000 description 1
- 241001497770 Melocanna baccifera Species 0.000 description 1
- 238000009825 accumulation Methods 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 230000003044 adaptive effect Effects 0.000 description 1
- 239000000654 additive Substances 0.000 description 1
- 230000000996 additive effect Effects 0.000 description 1
- 230000004075 alteration Effects 0.000 description 1
- 230000000712 assembly Effects 0.000 description 1
- 238000000429 assembly Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 238000000354 decomposition reaction Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000004880 explosion Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000002452 interceptive effect Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000003062 neural network model Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000003672 processing method Methods 0.000 description 1
- 230000004044 response Effects 0.000 description 1
- 238000005204 segregation Methods 0.000 description 1
- 230000003068 static effect Effects 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/253—Fusion techniques of extracted features
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/478—Contour-based spectral representations or scale-space representations, e.g. by Fourier analysis, wavelet analysis or curvature scale-space [CSS]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/20—Movements or behaviour, e.g. gesture recognition
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/27—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique
- G10L25/30—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique using neural networks
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L2021/02087—Noise filtering the noise being separate speech, e.g. cocktail party
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Health & Medical Sciences (AREA)
- Multimedia (AREA)
- General Physics & Mathematics (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Artificial Intelligence (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Acoustics & Sound (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Signal Processing (AREA)
- Biophysics (AREA)
- Software Systems (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- Quality & Reliability (AREA)
- Molecular Biology (AREA)
- Evolutionary Biology (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Social Psychology (AREA)
- Psychiatry (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Image Analysis (AREA)
- Telephonic Communication Services (AREA)
Abstract
The present disclosure provides a multi-modal speech separation method and system.
The method includes: receiving mixed speech of the speakers and facial visual information of
the speakers; performing face detection by using a Dlib library to obtain a quantity of
speakers; processing the above information to obtain a complex spectrogram and face images
of the speakers and transmitting the complex spectrogram and the face images to a
multi-modal speech separation model, and dynamically adjusting a structure of the model
according to the quantity of the speakers, where during training of the multi-modal speech
separation model, a complex ideal ratio mask (cIRM) is used as training target, where the
cIRM is defined as a ratio of a spectrogram of a clean speech to a spectrogram of the mixed
speech in a complex domain, and is composed of a real part and an imaginary part and
includes an amplitude and phase information of the speech, and the multi-modal speech
separation model outputs complex time-frequency masks for a quantity of faces; and
performing complex multiplication on the outputted masks and the spectrogram of the mixed
speech to obtain the spectrogram of the clean speech, and performing short-time Fourier
transform (STFT) on the spectrogram of the clean speech to obtain a time-domain signal of
the clean speech, thereby completing speech separation. The disclosed model is more
applicable to most application scenarios.
1/2
Mixed speech
ST FT-- - ---- --- --- - - --- ------ --- - -
Muli
Speaker
OeV+separation
i
FIG. I
CNN layers
Segmentation TCN layers
Audio-visual
Audio feature fusion
extraction
FC l ayers
Weight sharing
CNN layers
Visual feature
extraction
FIG. 2
Description
1/2
Mixed speech
Muli
i OeV+separation Speaker
FIG. I CNN layers Segmentation TCN layers
Audio-visual Audio feature fusion extraction
FC l ayers Weight sharing
CNN layers Visual feature extraction
FIG. 2
The present disclosure belongs to the technical field of speech separation, and in particular, to a multi-modal speech separation method and system.
The description in this section merely provides background information related to the present disclosure and does not necessarily constitute the prior art. In our daily life, we often need to be exposed to a variety of mixed speech, and person-to-person mixed speech is most frequently required to be processed. In an environment where a plurality of pieces of speech are mixed, we have the ability to focus on the speech of one person and ignore the speech of others and environmental noise. This phenomenon is referred to as the cocktail party effect. With the powerful speech signal processing capability, the human auditory system can easily separate the mixed speech. With the intelligence of life, speech separation technology has played an important role in various speech interaction devices. However, for computers, how to efficiently achieve speech separation has always been a difficult problem. At present, the speech separation technology is widely used in our life. For example, the speech separation technology is added to the front of speech recognition to separate the speech of the target speaker from other interfering speech, thereby improving the robustness of the speech recognition system. It is precisely because the speech separation technology can facilitate the subsequent speech signal processing, more and more people are focusing on the speech separation. In the past few decades, various speech separation algorithms have been proposed, which have been proven to effectively improve the performance of speech separation. Even so, the speech separation technology still has more development prospects. Among the proposed speech separation algorithms, most of them only use audio feature information for speech separation. There are some conventional methods, such as the independent component analysis (ICA)-based method, the computational auditory scene analysis (CASA)-based method, the Gaussian mixture model (GMM)-based method, and so on. ICA is a calculation method used to separate a plurality of signals into additive subcomponents, so as to realize the separation of speech signals by searching for statistically independent and non-Gaussian components in a multi-dimensional array, thereby realizing rapid analysis and processing of the speech signals. Therefore, the ICA is widely used in blind source separation [Blind equalisation using approximate maximum likelihood source separation][Adaptive blind source separation with HRTFs beamforming preprocessing][Convolutive Blind Source Separation Applied to the Wireless Communication]. The CASA uses the computer technology to perform modeling on the processing of auditory signals by using the computer imitating the human, so as to imitate the human perceiving, processing, and interpreting speech from complex mixed speech sources. In the reference [Speech segregation based on speech localization], the ideal binary mask (IBM) is combined with the CASA method to construct a new speech separation model, improving the intelligibility of the separated speech. The GMM is a clustering algorithm that uses Gaussian distribution as a parameter model and has been widely used in single-channel speech separation methods. The reference [Soft Mask Methods for Single-Channel Speaker Separation] provides a method for resolving single-channel speech separation by using the GMM, which is to learn the parameters of the GMM by using the expectation maximization (EM) algorithm. But the method still has shortcomings. It is difficult to choose the order of the source distribution, and the method relies heavily on initialization, which is very complicated to implement. With the rapid development of deep learning, some excellent algorithms have been proposed, such as convolutional neural network (CNN), recurrent neural network (RNN), and so on. Artificial neural networks have greatly improved the performance of supervised learning tasks due to their strong nonlinear mapping capabilities, and is concerned by more and more people. At present, most of the speech separation algorithms based on deep learning that have been proposed use time-frequency decomposition techniques such as short-time Fourier transform (STFT) in the data preprocessing process to convert speech into spectrogram. However, in recent years, there has been end-to-end speech separation methods that directly uses the mixed speech signal as the input. The reference [Joint Optimization of Masks and Deep Recurrent Neural Networks for Monaural Source Separation] proposed a speech separation method based on the deep CNN, which achieved more robust separation performance byjointly optimizing the time-frequency masking of multi-target sources. Since the conventional RNN has the vanishing gradient problem, an effective solution is to introduce a gating mechanism to control the accumulation speed of information, and long short-term memory (LSTM) is one of the typical representatives. In the literature [Long short-term memory for speaker generalization in supervised speech separation], the LSTM network is used instead of the RNN network in the speech separation algorithm, and it turns out that the separation performance has been improved. In addition, unlike the use of the speech spectrogram as input, the literature [TASNET: TIME-DOMAIN AUDIO SEPARATION NETWORK FOR REAL-TIME, SINGLE-CHANNEL SPEECH SEPARATION] proposed a model named TasNet for end-to-end speech separation in the time domain, which uses speech timing signals as inputs, combines one-dimensional convolution with LSTM, and finally directly outputs clean speech signals. Recently, it has been found that the use of a temporal convolutional network (TCN) is very effective in sequence modeling, and one-dimensional time series signals can directly use TCN to extract features, so that some end-to-end speech separation methods based on the TCN have been proposed. The literature [FurcaNext: End-to-End Monaural Speech Separation with Dynamic Gated Dilated Temporal Convolutional Networks] proposed four improved models for monaural end-to-end speech separation based on the TCN on the basis of TasNet, in which the FurcaPy model uses the multi-stream TCN to extract audio features, while the other three models directly modify the internal structure of the TCN to improve performance in speech separation tasks. The previous methods only use audio information. However, in the process of human speech perception, visual information is usually used automatically or unconsciously [Visual Speech Recognition: Lip Segmentation and Mapping]. When listening to a speaker, people will not only focus on the speech of the speaker but also look at the lip of the speaker to better understand what the speaker says. In addition, the permutation problem also exists when only audio information is used for speech separation. With the continuous development of smart devices, it has become increasingly more convenient to obtain the visual information of the speaker. Visual information can be used as a supplement to the speech separation model. Since some phonemes are easier to distinguish visually [CochleaNet: A robust language-independent audio-visual model for real-time speech enhancement], and when a speaker is in a silent state, the visual information can better reflect the situation. Therefore, compared to using only audio information, the combination of visual and audio information avoids the one-sidedness and uncertainty of obtaining information in a single mode, thereby improving the performance of speech separation. In recent years, many people have made full use of visual information and proposed some audio-visual fusion speech separation algorithms. In 2018, Google proposed a deep learning-based audio-visual fusion speech separation algorithm [Looking to Listen at the Cocktail Party: A Speaker-Independent
Audio-Visual Model for Speech Separation], which is to combine the CNN and the LSTM network to extract audio and visual features, use complex ideal ratio masks (cIRM) as the training target, and multiply the cIRM by mixed audio spectrogram to obtain clean speaker speech. The model achieves speaker independence, and the performance is significantly improved compared to the method that only use audio information. Due to too many network parameters, overfitting will inevitably occur during the training. In order to solve this problem, the literature [Two Stage Audio-Video Speech Separation using Multimodal Convolutional Neural Networks] proposed a system having two AV models to separate the mixed speech. The two models were trained separately. After the first AV model was trained, the output and visual features would be used to train the second AV model. In addition, the literature [Audio-Visual Deep Clustering for Speech Separation] used an unsupervised learning method and proposed a speech separation model (AVDC) with two feature fusions. The use of visual information can better help the clustering of time-frequency units. Compared with the speech separation model that only uses audio information, AVDC has a better separation effect and resolves the problem of permutation problem across frames. In addition to the audio frequency domain processing methods, visual information is incorporated on the basis of end-to-end speech separation by some methods. The literature
[TIME DOMAIN AUDIO VISUAL SPEECH SEPARATION] combined visual information on the basis of TasNet to realize multi-model learning. Experiments show that the model has better effects than the method using only audio information. Based on the above literature, only one feature extractor is used to extract the audio feature of the full band, that is, the feature extractor used on each frequency is the same, and the obtained data has undesirable effects. In addition, the number of existing speakers is fixed when the network parameters are designed, that is, the model is static. A fixed number of speakers are required during training and testing, and the number of inputs cannot be flexibly changed. Therefore, the existing speech separation technology needs further improvement.
In order to overcome the above shortcomings of the prior art, the present disclosure provides a multi-modal speech separation method, so that the selected model is more effective than other models. To achieve the foregoing objective, one or more embodiments of the present disclosure provide the following technical solutions:
In a first aspect, a multi-modal speech separation method is disclosed, including: receiving mixed speech of speakers and facial visual information of the speakers, and obtaining a quantity of speakers by means of face detection; preprocessing the above data to obtain a complex spectrogram of the mixed speech and face images of the speakers and transmitting the complex spectrogram and the face images to a multi-modal speech separation model, and dynamically adjusting the structure of model according to the quantity of the speakers, where during training of the multi-modal speech separation model, a complex ideal ratio mask (cIRM) is used as a training target, where the cIRM is defined as a ratio of a spectrogram of a clean speech to the spectrogram of mixed speech in a complex domain, and is composed of a real part and an imaginary part and includes an amplitude and phase information of the speech, and the multi-modal speech separation model is configured to output a same quantity of cIRMs as the speakers; and performing complex multiplication on the outputted cIRMs and the spectrogram of the mixed speech to obtain the spectrogram of the clean speech, and performing inverse short-time Fourier transform (iSTFT) on the spectrogram of the clean speech to obtain a time-domain signal of the clean speech, thereby completing speech separation. In a further technical solution, the mixed speech may be regarded as a result of adding the clean speech of a plurality of speakers. The mixed speech signal in the time domain is converted, by means of short-time Fourier transform (STFT), to the complex spectrogram that may be used as an input of the speech separation model. In a further technical solution, before each instance is inputted into the multi-modal speech separation model, the structure of the model is dynamically adjusted according to the quantity of speakers. In a further technical solution, the multi-modal speech separation model is composed of an audio feature extraction network, a visual feature extraction network, and an audio-visual fusion network. The audio feature extraction network is configured to extract high-frequency audio features and low-frequency audio features by using different convolutional neural networks (CNN), fuse the low-frequency audio feature and the high-frequency audio feature to realize a first stage of fusion, and then continue to extract audio features by using temporal convolutional networks (TCN). The visual feature extraction network is configured to extract visual features for the inputted face images by using a plurality of convolutional layers, and insert a hole into each convolution kernel to increase a size of the each convolution kernel, thereby increasing a receptive field. The audio-visual fusion network is configured to fuse the audio feature obtained by the audio feature extraction network and the visual feature obtained by the visual feature extraction network to obtain audio-visual fused features, thereby realizing a second stage of fusion of the features. Preferably, the extracting high-frequency audio features and low-frequency audio features by using different feature extractors specifically includes: transforming the time-domain signal of the mixed speech into complex spectrogram by using STFT, and then segmenting the complex spectrogram into a high-frequency part and a low-frequency part in frequency dimension; extracting the low-frequency audio feature and the high-frequency audio feature by using a two-stream CNN, where each stream includes two convolutional layers, where different dilation parameters are used for network layers for extracting the high-frequency features and network layers for extracting the low-frequency features; and fusing the high-frequency audio feature and the low-frequency audio feature to realize the first stage of fusion of the features. Preferably, one-dimensional convolutional network layers in the TCN are modified into two-dimensional convolutional network layers, so that the TCN is capable of processing output data of the audio feature extraction network. Preferably, an output of the visual feature extraction network is up-sampled to compensate for a sampling rate difference between an audio signal and a video signal. The audio feature and the visual feature are fused to realize the second stage of fusion of the features. In a further technical solution, the multi-modal speech separation model inputs audio-visual fused features to a fully connected layer. The fully connected layer outputs cIRMs for a quantity of speakers. Each cIRM corresponds to one speaker, and a sequence of the speakers corresponding to the masks is the same as a sequence of the speakers in the visual feature extraction network. In a second aspect, a multi-modal speech separation system is disclosed, including: a data receiving module, configured to receive mixed speech of speakers and facial visual information of the speakers; a multi-modal speech separation model processing module, configured to process the above information to obtain a complex spectrogram and face images and transmit the complex spectrogram and the face images to a multi-modal speech separation model, and dynamically adjust a structure of the model according to a quantity of the speakers, where during training of the multi-modal speech separation model, complex time-frequency masks are used as a training target, and the multi-modal speech separation model outputs complex time-frequency masks for a quantity of faces; and a speech separation module, configured to perform complex multiplication on the outputted masks and the spectrogram of the mixed speech to obtain the spectrogram of the clean speech, and perform STFT on the spectrogram of the clean speech to obtain a time-domain signal of the clean speech, thereby completing speech separation. The foregoing one or more technical solutions have the following beneficial effects: In the technical solution of the present disclosure, a multi-modal speech separation system is disclosed to resolve the cocktail party problem, as shown in FIG. 1. The human-based auditory system has a powerful speech separation capability. Therefore, the speech separation system of the technical solution of the present disclosure refers to the relevant physiological characteristics of people. The part of the human ear that receives speech is the cochlea, which can map speech of different frequencies to different positions on the basilar membrane. Based on the characteristic of the cochlea, a multi-stream convolutional network is used in the audio feature extraction network to simulate feature extraction filters of the cochlea in different frequency ranges, so as to extract the features of a high frequency part and a low frequency part of the speech respectively. In addition, the model of the technical solution of the present disclosure includes two stages of fusions of features: a fusion of high-frequency audio feature and low-frequency audio feature and a fusion of audio feature and visual feature. The improved TCN is used to process the fused features of high-frequency speech and low-frequency speech, and visual information is added to the speech separation model to improve the performance of speech separation. For each instance, the network structure of the model and the quantity of finally outputted speech are determined by a quantity of speakers in the video. Therefore, the model is flexible and suitable for speech separation of any quantity of speakers. Since different parts of the basilar membrane in the cochlea may process speech of different frequencies, the technical solution of the present disclosure uses two different feature extractors to extract the features of the high-frequency part and the features of the low-frequency part of the speech respectively.
According to the technical solution of the present disclosure, the speech separation model is combined with the face detection, so that the system can recognize how many speakers are in the video, thereby dynamically adjusting the structure of the model. In view of the permutation problem in the prior art that only uses audio information for speech separation, the technical solution of the present disclosure combines audio and video to fix the sequence of a plurality of outputs, thereby resolving the permutation problem. In addition, the visual information may be used as a supplementary part of the speech separation system, so as to further improve the performance. In view of the prior art using the RNN to perform time series modeling, a large quantity of parameters are required, a long time is needed, and the problems of gradient exploding and gradient vanishing may also exist. The technical solution of the present disclosure uses the TCN instead of a long short-term memory (LSTM). The advantages of the TCN mainly include parallelism, a flexible receptive field, a stable gradient, and lower memory. According to the technical solution of the present disclosure, SDR is selected as the evaluation index, the model of the technical solution of the present disclosure is compared with other superior models through experiments, and the feasibility of the model is verified through some auxiliary experiments. The additional aspects and advantages of the present disclosure will be set forth in part in the description below, parts of which will become apparent from the description below, or will be understood by the practice of the present disclosure.
The accompanying drawings constituting a part of the present disclosure are used to provide further understanding of the present disclosure. Exemplary embodiments of the present disclosure and descriptions thereof are used to explain the present disclosure, and do not constitute an improper limitation to the present disclosure. FIG. 1 is a diagram of a speech separation system according to an embodiment of the present disclosure. FIG. 2 is a structure diagram of a multi-modal speech separation model according to an embodiment of the present disclosure. FIG. 3 is an improved diagram of a temporal convolutional network (TCN) according to an embodiment of the present disclosure.
It should be noted that, the following detailed descriptions are all exemplary, and are intended to provide further descriptions of the present disclosure. Unless otherwise specified, all technical and scientific terms used herein have the same meaning as commonly understood by a person of ordinary skill in the technical field to which the present disclosure belongs. It should be noted that terms used herein are only for describing specific implementations and are not intended to limit exemplary implementations according to the present disclosure. As used herein, the singular form is intended to include the plural form, unless the context clearly indicates otherwise. In addition, it should be further understood that terms "include" and/or "comprise" used in this specification indicate that there are features, steps, operations, devices, assemblies, and/or combinations thereof. The embodiments in the present disclosure and features in the embodiments may be mutually combined in case that no conflict occurs. Overall conception: In an environment where a plurality of pieces of speech are mixed, it is often necessary to separate the speech of interest to facilitate subsequent speech processing. However, for computers, how to efficiently achieve speech separation has always been a relatively major problem. In the technical solution of the present disclosure, a deep learning-based multi-modal speech separation model is provided to resolve the cocktail party problem. The model makes full use of audio information and visual information of speakers, and adopts a two-stage feature fusion policy: a fusion of high-frequency audio feature and low-frequency audio feature and a fusion of audio feature and visual feature. The multi-stream convolution is used to process the high-frequency audio and the low-frequency audio separately, and connect the outputs of the high-frequency audio feature and the low-frequency audio feature as the input of a temporal convolutional network (TCN) to achieve audio feature extraction. The obtained audio feature is combined with the visual feature outputted by a dilated convolutional layer, and the speech separation is completed through the fully connected layers. In addition, at the data preprocessing stage, a Dlib library is used to detect a quantity of speakers in the video and dynamically adjust the network structure, so as to automatically determine how many clean outputs need to be generated. The data set used in the present disclosure is GRID. In order to make full use of the phase information of the speech, the cIRM is selected as the training target of the model. Through a series of experiments, it is proved that the performance of the disclosed model is superior to that of the model using other methods. In addition, additional experiments prove that the model of the present disclosure has a faster training speed than the model proposed by Google without affecting performance, and it is verified that visual information is indeed helpful for improving the performance of speech separation. Embodiment I As shown in FIG. 1, this embodiment discloses a multi-modal speech separation method, including: receiving mixed speech of speakers and facial visual information of the speakers, and obtaining a quantity of speakers by means of face detection; processing the above information to obtain a complex spectrogram and face images and transmitting the complex spectrogram and the face images to a multi-modal speech separation model, and dynamically adjusting a structure of the model according to the quantity of the speakers, where during training of the multi-modal speech separation model, a complex ideal ratio mask (cIRM) is used as a training target, where the cIRM is defined as a ratio of a spectrogram of clean speech to a spectrogram of the mixed speech in complex domain, and is composed of a real part and an imaginary part and includes an amplitude and phase information of the speech, and the multi-modal speech separation model is configured to output cIRMs for a quantity of speakers; and performing complex multiplication on the outputted cIRMs and the spectrogram of the mixed speech to obtain the spectrogram of the clean speech, and performing short-time Fourier transform (STFT) on the spectrogram of the clean speech to obtain a time-domain signal of the clean speech, thereby completing speech separation. In a specific implementation example, speech separation may be deemed as mixed speech x(t) composed of audio of a plurality of speakers, and the audio of the speakers are estimated as si(t), ..., and sc(t). The mixed speech x(t) may be deemed to be obtained by mixing the clean audio of C speakers: C x(t) si(t) i=1
The mixed speech signal in the time domain is converted, by means of short-time Fourier transform (STFT), to a complex spectrogram X that may be used as an input of the speech separation model, which includes a real part and an imaginary part. For a visual stream, the video is first converted to video frames, then Dlib is used to determine a quantity C of human faces in the video, and all faces are extracted as the input of the model.
During the training of the model, time-frequency masks are used as the training target, that is, for each speaker, the model will predict a time-frequency mask. It can be seen from the literatures [On training targets for supervised speech separation][Supervised Speech Separation Based on Deep Learning: An Overview] that compared with directly prediction of the spectrogram or time-domain waveforms of a clean speech, the mask is used as the training target, so that the speech separation system can have better performance. In addition, the literature [The importance of phase in speech enhancement] pointed out that the phase information of the speech is also very helpful for speech separation. The time-frequency mask used is complex ideal ratio mask. The cIRM is defined in the complex domain as a ratio of the spectrogram of the clean speech and the spectrogram of the mixed speech, consists of the real part and the imaginary part [Complex Ratio Masking for Monaural Speech Separation], and includes the amplitude and the phase information of the speech. Since the faces of C speakers are detected, the model outputs C cIRMs, and Mi, M 2 , ... , and Mc are used to represent the cIRM corresponding to each speaker. Then, complex multiplication is performed on the cIRMs and the spectrogram of the mixed speech to obtain the spectrogram of the clean speech. Y,,(t) = M,, * X, n = 1, 2, . . , C where '*' represents the complex multiplication, X represents the complex spectrogram obtained by STFT conversion of the inputted mixed speech signal. Finally, the spectrograms are obtained by means of inverse short-time Fourier transform (iSTFT) to obtain the time-domain signal of the clean audio, thereby completing the speech separation. The multi-modal speech separation model is to be described in detail below, which is composed of an audio feature extraction network, a visual feature extraction network, and an audio-visual feature fusion network. The model structure is shown in FIG. 2. Taking two speakers as an example, the model is to be described in detail below. Regarding the audio feature extraction network, the features of the cochlea are used in the design of the audio feature extraction network for reference. Since humans can recognize and separate speech of different frequencies through the auditory system, the cocktail effect problem can be easily resolved. The cochlea plays an important role in audio processing, which can map speech of different frequencies to different locations of the basement membrane. The basement membrane at the bottom of the cochlea may process high-frequency speech, and the basement membrane at the top of the cochlea may process low-frequency speech. In addition, in the literature [TasNet: TIME-DOMAIN AUDIO SEPARATION NETWORK FOR REAL-TIME, SINGLE-CHANNEL SPEECH
SEPARATION], the author designed and trained a CNN-based feature extraction filter to make the network find a filter suitable for speech separation, and found from the frequency response diagram of the features that the frequency resolutions of the filter at the low frequency and the high frequency are different. Specifically, the low frequency part has a high resolution and the high frequency part has a low resolution, which indicates that the optimal filters for extracting the audio features at the low frequency and the high frequency are different in property. Therefore, high-frequency audio feature and low-frequency audio feature are to be extracted by using different convolutional neural networks (CNN). First, the time-domain signal of the mixed speech is transformed into a complex spectrogram by using an STFT, and then the complex spectrogram is segmented into a high-frequency part and a low-frequency part in a frequency dimension. Inspired by the cochlea, the low-frequency audio features and the high-frequency audio features are extracted by using a two-stream CNN. Each stream includes two convolutional layers. Each convolutional layer includes 32 convolution kernels having a size of 25x5. For the network layer that extracts high-frequency features, a dilation parameter of the first convolutional layer is set to 1x1, and a dilation parameter of the second convolutional layer is set to 2x1. For the network layer that extracts low-frequency features, the dilation parameter is set to 1x1 uniformly. According to the technical solution of the present disclosure, the low-frequency audio feature and the high-frequency audio feature are fused to realize a first stage of fusion, and then the TCN is used to continue to extract the features. The TCN not only can process data in parallel, effectively prevent gradient explosions, but also has strong flexibility [An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling]. Therefore, increasingly more people consider using the TCN instead of the LSTM to process sequence data. Since the conventional TCN uses one-dimensional convolution internally, only one-dimensional data can be used as an input, and therefore the conventional TCN is more commonly used in end-to-end speech separation models. Since the original speech signal is not directly used as the input, but transformed into a complex spectrogram by the STFT, this requires the two-dimensional convolution to process the data. The one-dimensional convolution in the TCN is modified into two-dimensional convolution, so that the TCN is capable of processing the output data of the previous network layer normally. The internal structure improvement of the TCN is shown in FIG. 3. A total of 8 layers of the TCN are used, and batch normalization is used after each layer of the TCN to prevent overfitting. Regarding the visual feature extraction network, visual features are extracted for the inputted face images by using 6 convolutional layers, and a hole is inserted into each convolution kernel to increase a size of the each convolution kernel, thereby increasing a receptive field. The parameters of convolution kernels included in the convolutional layer are respectively 32, 48, 64, 128, 256, and 256, and a size of each convolution kernel is 5x1. Similar to the audio feature extraction network, batch normalization is used after each convolutional layer to prevent overfitting. Finally, in order to compensate for the sampling rate difference between the audio signal and the video signal, the output of the visual feature extraction network is up-sampled. It is to be noted that since the mixed speech is composed of audio of at least two persons, the faces of a plurality of speakers need to be inputted during the visual feature extraction. In order to achieve speaker independence, in the process of extracting the visual features of different speakers, weight sharing is realized. That is to say, the network having the same parameters is used to extract face features, so that the visual feature extraction network has the generalization performance. Regarding the audio and visual features fusion network, after the mixed speech and the visual information of speakers respectively pass through the audio feature extraction network and the visual feature extraction network, the audio feature and the visual feature are obtained. Then the audio feature and the visual feature are fused to obtain audio-visual fused features, thereby realizing a second stage of fusion of the features. Then the fused features are inputted into fully connected (FC) layers. A non-linear ReLU activation function is used in FC layers. 3 FC layers are used to process the fused features, and each FC layer includes 500 units. The FC layer outputs C cIRMs. Each mask corresponds to one speaker. The sequence of the masks corresponding to the speakers is the same as the sequence of the speakers corresponding to the fusion of the visual features in the audio and visual feature fusion process, so as to resolve the permutation problem caused the label arrangement during the speech separation using only audio information. Complex multiplication is performed on the cIRMs and the complex spectrogram of the mixed speech to obtain the complex spectrogram of the clean speech of the corresponding speaker. Finally, the complex spectrogram corresponding to each person is transformed into a clean speech signal by means of the iSTFT. Embodiment II This embodiment is intended to provide a computing device, including a memory, a processor, and a computer program stored in the memory and executable by the processor, where when the processor executes the program, steps of the foregoing method are performed. Embodiment III This embodiment is intended to provide a computer-readable storage medium. The computer-readable storage medium stores a computer program, where when the program is executed by a processor, steps of the foregoing method are performed. Embodiment IV This embodiment is intended to provide a multi-modal speech separation system, including: a data receiving module, configured to receive mixed speech of speakers and facial visual information of the speaker; a multi-modal speech separation model preprocessing module, configured to preprocess the above data to obtain a complex spectrogram and face images and transmit the complex spectrogram and the face image to a multi-modal speech separation model, and dynamically adjust a structure of the model according to a quantity of the speakers, where during training of the multi-modal speech separation model, a cIRM is used as a training target, where the cIRM is composed of a real part and an imaginary part and includes an amplitude and phase information of the speech, and the multi-modal speech separation model is configured to output cIRMs for a quantity of faces; and a speech separation module, configured to perform complex multiplication on outputted time-frequency masks and a spectrogram of the mixed speech to obtain spectrogram of clean speech, and perform STFT on the spectrogram of the clean speech to obtain a time-domain signal of the clean speech, thereby completing speech separation. The steps involved in the devices of the above embodiments II, III, and IV correspond to the method embodiment I. For the specific implementation, reference may be made to the relevant description part of Embodiment I. The term "computer-readable storage medium" should be understood as a single medium or a plurality of media including one or more instruction sets, and should also be understood to include any medium that can store, encode, or carry an instruction set executed by the processor and enable the processor to execute any method in the present disclosure. Experiment and result: This part introduces the data set used and the setting of the parameters in the process of building the experiment, and then gives experimental results and analyzes the results. In order to facilitate comparison with the model results in other literatures, a quantity of speakers is set to 2. That is to say, the mixed speech includes two speakers, and the model finally outputs two speech signals. Regarding the data set: The multi-modal speech separation model provided in the present disclosure is trained by using a GRID data set. The GRID data set records facial video and audio information of 34 speakers. For each speaker, the data set includes 1000 facial video and corresponding audio. A duration of each video is 3s, and a frame rate of the video is 25 FPS. A mixed speech data set is further constructed by referring to [Audio-Visual Deep Clustering for Speech Separation]. Since data information corresponding to some speakers is incomplete, all data of these speakers is deleted. After preprocessing, the finally selected data set includes 17 male speakers and 15 female speakers. The model selects data from two male speakers and two female speakers to build a validation set, selects data from two other male speakers and two female speakers to build a test set, and selects the rest of the data to build a training set. All audio is down-sampled to 16 kHz. During construction of mixed speech, the audio of different speakers is randomly selected, and then the clean audio is fused to obtain the mixed speech. In this experiment, the mixed speech of two persons was used, and a total of 41847 mixed speech was finally obtained. Data preprocessing: Complex spectrograms of all mixed audio are obtained by using the STFT. A length of a Hamming window is 25 ins, a sliding distance is 10 ins, and a window size of fast Fourier transform (FFT) is 512. Since the obtained complex spectrogram includes a real part and an imaginary part, a magnitude of the spectrogram that is finally obtained by means of the STFT are 257x298x2. Then the complex spectrogram is divided into a high frequency part and a low frequency part in the frequency dimension. After many attempts, it is found that it is the most appropriate when the segmentation point in the frequency dimension is 180. Specifically, the magnitude of the low-frequency audio feature is 180x298x2, and the magnitude of the high-frequency audio feature is 77x298x2. For video data, since each video is 25 FPS, a total of 3s, after the video is transformed into video frames, 75 pictures may be obtained. OpenCV and Dlib are used to obtain a quantity of faces, locate a face area of each frame, and extract the face area. A size of each face picture that is finally obtained is 160x160. Experimental setup: In the experiment, a deep learning framework used is Keras, and the NVIDIA RTX 2080Ti graphic card is used to train the model. The network has trained 300 epochs and uses the Adam optimizer. An initial learning rate is set to 0.00001. In order to prevent overfitting, when every 100 epochs are trained, the learning rate is reduced to one-tenth of the previous epoch. Experimental result: An evaluation index that is used is an SDR. Since different mixed speech corresponds to different SDRs, the quality of the model cannot be determined simply by comparing the SDR of the output clean speech. For the convenience of comparison, the SDR of the finally outputted speech of the model is subtracted from the SDR of the mixed speech, and an actual improved SDR may be obtained, that is, ASDR. When there are only two speakers, the ASDR results using different methods are compared and analyzed. The inputted mixed speech is divided into three situations: male-male mixed speech, male-female mixed speech, and female-female mixed speech. When the inputs are respectively the male-male mixed speech, the male-female mixed speech, the female-female mixed speech, the performance of the speech separation model using different methods is obtained, and the results are shown in Table 1. It can be seen that compared with other speech separation methods, the method of the present disclosure has superior performance. Since the frequencies of male audio and female audio are quite different, the effect of speech separation on mixed audio of the opposite sex is the best, thereby achieving an increase of 11.14 dB in the model of the present disclosure. When the input is mixed speech of the same sex, compared with the mixed speech of the opposite sex, the separation effect of all models is reduced. However, the model of the present disclosure also performs better than other models, achieving an increase of 6.31 dB in the male-male mixed speech and an increase of 8.47 dB in the female-female mixed speech. Table 1 Comparison of the speech separation results of different models when the mixed speech includes only speech of two speakers
ASDR (dB) Method Male-male mixed Male-female mixed Female-female mixed speech speech speech AV-Match 4.93 9.80 6.96 AVDC 6.18 10.16 8.32 uPIT 4.68 9.56 4.92 Model of the 6.31 11.14 8.47 present disclosure . 1 8
In addition, the model of the present disclosure also resolves the problems of a large quantity of parameters and slow training time that often appear in neural network models. During the feature extraction of speech, the TCN is used instead of the LSTM for time series modeling, which greatly reduces training parameters and the training time. Since the model using the LSTM is similar to the Google model, the same data set is initially used to compare the model of the present disclosure with the Google model. During the experiment, it was found that the AVSpeech data set proposed by Google was too large and the training time was very long, which was not conducive to the experiment. Therefore, the GRID data set was used to train the model proposed by Google to be compared with the model of the present disclosure. In order to save time costs and ensure fairness, the epoch of the training parameters of the two models is set to 100. At this point, although the loss of the two models has not reached the lowest point, the comparison of the two models is not affected. The comparison results are shown in Table 2. It can be seen that under the same conditions, compared with the Google model, the model of the present disclosure can greatly save the training time, and the separation performance is not affected, but has been improved. Therefore, the model of the present disclosure is more applicable to most application scenarios. Table 2 Comparison of results between the model of the present disclosure and the Google model
Method Training time (100 epochs) ASDR (100 epochs) Model provided by Google About 400 h 9.43 Model of the present disclosure About 72 h 9.86
As described above, in this application, high-frequency audio feature and low-frequency audio feature in the model audio feature network are extracted by using a two-stream CNN. An auxiliary experiment is designed to prove that using a plurality of streams to extract features indeed facilitates speech separation. In a new model, a high-frequency feature extraction network is deleted, and the mixed spectrogram may be directly fed to a low-frequency feature extraction network without segmentation. The comparison results are shown in Table 3. It can be seen from the table that the features extracted by using a multi-stream CNN may be 1.34 dB higher than that extracted by using a single-stream CNN. Therefore, the best method is to use different networks to extract high-frequency feature and low-frequency feature separately, because ideal feature extractors corresponding to different frequencies of the speech are different. Table 3 Impact of the two-stream CNN and the single-stream CNN on the results of extracting audio features
Method ASDR (dB) Audio feature extraction by using the two-stream CNN 11.14 Audio feature extraction by using the single-stream CNN 9.65
In order to observe whether visual information is helpful to the performance of speech separation, the speech separation model that only uses the speech is compared with the audio-visual fusion model. Specifically, the technical solution of the present disclosure uses only the speech stream in the proposed model and deletes the visual feature extraction network, and the others remain unchanged, so that a speech separation model that only uses the mixed speech as the input is obtained. The comparison results are shown in Table 3. It can be seen from the table that adding visual feature information has a certain effect on the improvement of speech separation performance.
Table 4 Comparison of results between the model using only the audio and the audio-visual feature fusion model
Method ASDR (dB) Speech separation using only the audio 10.32
Speech separation of audio-visual fusion 11.14
The technical solution of the present disclosure provides a new audio-visual fusion speech separation model. The model combines audio features and visual features, imitates the physiological characteristics of the cochlea in the process of audio feature extraction, extracts the high-frequency features and low-frequency features of the speech by using different networks, and can automatically determine a quantity of outputted models by using the face detector. It can be seen from the experimental results that when the same data set is used, the model of the present disclosure has better performance than several models recently provided. In addition, the feasibility of the ideas proposed in the technical solution of the present disclosure is verified by means of some experiments that have been conducted. In the subsequent technical solutions, the speech separation technology may be applied to more complex scenarios, and is applied to speech enhancement to achieve background noise suppression. A person skilled in the art should understand that the modules or steps in the present disclosure may be implemented by using a general-purpose computer apparatus. Optionally, they may be implemented by using program code executable by a computing apparatus, so that they may be stored in a storage apparatus and executed by the computing apparatus. Alternatively, the modules or steps are respectively manufactured into various integrated circuit modules, or a plurality of modules or steps thereof are manufactured into a single integrated circuit module. The present disclosure is not limited to any specific combination of hardware and software. The foregoing descriptions are merely exemplary embodiments of the present disclosure, but not intended to limit the present disclosure. Those skilled in the art may make various alterations and variations to the present disclosure. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present disclosure shall fall within the protection scope of the present disclosure. The specific implementations of the present disclosure are described above with reference to the accompanying drawings, but are not intended to limit the protection scope of the present disclosure. A person skilled in the art should understand that various modifications or transformations may be made without creative efforts based on the technical solutions of the present disclosure, and such modifications or transformations shall fall within the protection scope of the present disclosure.
Claims (9)
1. A multi-modal speech separation method, comprising: receiving mixed speech of speakers and facial visual information of the speakers, and obtaining a quantity of the speakers by means of face detection; pre-processing the mixed speech and facial visual information of the speakers to obtain a complex spectrogram of the mixed speech and face images of the speakers, transmitting the spectrogram and the face images to a multi-modal speech separation model, and dynamically adjusting a structure of the model according to the quantity of the speakers, wherein the multi-modal speech separation model is configured to output complex ideal ratio masks (cIRM) for the quantity of speakers; and performing complex multiplication on the outputted cIRMs and the complex spectrogram of the mixed speech to obtain a spectrogram, and transforming the spectrogram to obtain a time-domain signal of a clean speech, thereby completing speech separation; wherein, the multi-modal speech separation model, comprising: an audio feature extraction network, being configured to extract a high-frequency audio feature and a low-frequency audio feature by using different convolutional neural networks (CNN), and fuse the low-frequency audio feature and the high-frequency audio feature to realize a first stage of fusion of the features, and then continue to extract an audio feature by using a temporal convolutional network (TCN); a visual feature extraction network, being configured to extract a visual feature for inputted face images by using a plurality of convolutional layers, and insert a hole into each convolution kernel to increase a size of the each convolution kernel, thereby increasing a receptive field; and an audio-visual feature fusion network, being configured to fuse the audio feature obtained by the audio feature extraction network and the visual feature obtained by the visual feature extraction network to obtain an audio-visual fusion feature, thereby realizing a second stage of fusion of the features.
2. The multi-modal speech separation method according to claim 1, wherein the multi modal speech separation model has a dynamic network structure, for each instance, the structure of the model is dynamically adjusted according to the quantity of the speakers detected at a data preprocessing stage, and the multi-modal speech separation model is applicable to any quantity of the speakers; and preferably, during training of the multi-modal speech separation model, each cIRM is used as a training target, wherein the cIRM is defined as a ratio of a spectrogram of the clean speech to the spectrogram of the mixed speech in a complex domain, and is composed of a real part and an imaginary part and comprises an amplitude and phase information of the speech.
3. The multi-modal speech separation method according to claim 1, wherein the extracting high-frequency audio feature and low-frequency audio feature by using different feature extractors specifically comprises: transforming the time-domain signal of the mixed speech into a complex spectrogram by using a short-time Fourier transform (STFT), and then segmenting the complex spectrogram into a high-frequency part and a low-frequency part in a frequency dimension; extracting the low-frequency audio feature and the high-frequency audio feature by using a two-stream CNN, wherein each stream comprises two convolutional layers, wherein different dilation parameters are used for network layers for extracting the high-frequency feature and network layers for extracting the low-frequency feature; and fusing the high-frequency audio feature and the low-frequency audio feature to realize the first stage of fusion of the features.
4. The multi-modal speech separation method according to claim 1, wherein a one dimensional convolutional network layer in the TCN is modified into a two-dimensional convolutional network layer, so that the TCN is capable of processing output data of the audio feature extraction network.
5. The multi-modal speech separation method according to claim 1, wherein an output of the visual feature extraction network is up-sampled to compensate for a sampling rate difference between an audio signal and a video signal; and the audio feature and the visual feature are fused to realize the second stage of fusion of the features.
6. The multi-modal speech separation method according to claim 1, wherein the multi modal speech separation model inputs the audio-visual fusion feature to fully connected layer, wherein the fully connected layer outputs cIRMs for a quantity of faces, wherein each mask corresponds to one speaker, and a sequence of the speakers corresponding to the masks is the same as a sequence of the speakers corresponding to connection of visual features during a fusion process of the audio-visual fusion feature.
7. A multi-modal speech separation system, comprising: a data receiving module, being configured to receive mixed speech of the speakers and facial visual information of the speakers, and obtaining a quantity of the speakers by means of face detection; a multi-modal speech separation model processing module, being configured to process the above information to obtain a complex spectrogram and face images of the speakers, transmit the complex spectrogram and the face images to a multi-modal speech separation model, and dynamically adjust a structure of the model according to a quantity of the speakers, wherein during training of the multi-modal speech separation model, a complex time-frequency mask is used as a training target, wherein the complex time-frequency mask is defined as a ratio of a spectrogram of a clean speech to the spectrogram of the mixed speech in complex domain, and is composed of a real part and an imaginary part and comprises amplitude and phase information of the speech; wherein, the multi-modal speech separation model outputs complex time-frequency masks for the quantity of the speakers, comprising: an audio feature extraction network, being configured to extract a high-frequency audio feature and a low-frequency audio feature by using different convolutional neural networks (CNN), and fuse the low-frequency audio feature and the high-frequency audio feature to realize a first stage of fusion of the features, and then continue to extract an audio feature by using a temporal convolutional network (TCN); a visual feature extraction network, being configured to extract a visual feature for inputted face images by using a plurality of convolutional layers, and insert a hole into each convolution kernel to increase a size of the each convolution kernel, thereby increasing a receptive field; and an audio-visual feature fusion network, being configured to fuse the audio feature obtained by the audio feature extraction network and the visual feature obtained by the visual feature extraction network to obtain an audio-visual fusion feature, thereby realizing a second stage of fusion of the features; and a speech separation module, being configured to perform complex multiplication on the outputted masks and the spectrogram of the mixed speech to obtain the spectrogram of the clean speech, and perform inverse short-time Fourier transform (iSTFT) on the spectrogram of the clean speech to obtain a time-domain signal of the clean speech, thereby completing speech separation.
8. A computing device, comprising a memory, a processor, and a computer program stored in the memory and executable by the processor, wherein when the processor executes the program, steps of the method according to any of claims 1 to 6 are performed.
9. A computer-readable storage medium storing a computer program therein, wherein when the program is executed by a processor, steps of the method according to any of claims 1 to 6 are performed.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN2021102717038 | 2021-03-12 | ||
CN202110271703.8A CN113035227B (en) | 2021-03-12 | 2021-03-12 | Multi-modal voice separation method and system |
Publications (2)
Publication Number | Publication Date |
---|---|
AU2022200439A1 AU2022200439A1 (en) | 2022-09-29 |
AU2022200439B2 true AU2022200439B2 (en) | 2022-10-20 |
Family
ID=76470471
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
AU2022200439A Active AU2022200439B2 (en) | 2021-03-12 | 2022-01-24 | Multi-modal speech separation method and system |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN113035227B (en) |
AU (1) | AU2022200439B2 (en) |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP7242903B2 (en) * | 2019-05-14 | 2023-03-20 | ドルビー ラボラトリーズ ライセンシング コーポレイション | Method and Apparatus for Utterance Source Separation Based on Convolutional Neural Networks |
CN113593601A (en) * | 2021-07-27 | 2021-11-02 | 哈尔滨理工大学 | Audio-visual multi-modal voice separation method based on deep learning |
CN115691538A (en) * | 2021-07-29 | 2023-02-03 | 华为技术有限公司 | Video processing method and electronic equipment |
CN115938385A (en) * | 2021-08-17 | 2023-04-07 | 中移(苏州)软件技术有限公司 | Voice separation method and device and storage medium |
CN114245280B (en) * | 2021-12-20 | 2023-06-23 | 清华大学深圳国际研究生院 | Scene self-adaptive hearing aid audio enhancement system based on neural network |
CN114446316B (en) * | 2022-01-27 | 2024-03-12 | 腾讯科技(深圳)有限公司 | Audio separation method, training method, device and equipment of audio separation model |
CN114598914A (en) * | 2022-02-17 | 2022-06-07 | 厦门快商通科技股份有限公司 | Human voice separation method based on video, terminal equipment and storage medium |
CN114566189B (en) * | 2022-04-28 | 2022-10-04 | 之江实验室 | Speech emotion recognition method and system based on three-dimensional depth feature fusion |
CN114743561A (en) * | 2022-05-06 | 2022-07-12 | 广州思信电子科技有限公司 | Voice separation device and method, storage medium and computer equipment |
CN115035907B (en) * | 2022-05-30 | 2023-03-17 | 中国科学院自动化研究所 | Target speaker separation system, device and storage medium |
CN117238311B (en) * | 2023-11-10 | 2024-01-30 | 深圳市齐奥通信技术有限公司 | Speech separation enhancement method and system in multi-sound source and noise environment |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019104229A1 (en) * | 2017-11-22 | 2019-05-31 | Google Llc | Audio-visual speech separation |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11373672B2 (en) * | 2016-06-14 | 2022-06-28 | The Trustees Of Columbia University In The City Of New York | Systems and methods for speech separation and neural decoding of attentional selection in multi-speaker environments |
CN108010514B (en) * | 2017-11-20 | 2021-09-10 | 四川大学 | Voice classification method based on deep neural network |
CN110931022B (en) * | 2019-11-19 | 2023-09-15 | 天津大学 | Voiceprint recognition method based on high-low frequency dynamic and static characteristics |
-
2021
- 2021-03-12 CN CN202110271703.8A patent/CN113035227B/en active Active
-
2022
- 2022-01-24 AU AU2022200439A patent/AU2022200439B2/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2019104229A1 (en) * | 2017-11-22 | 2019-05-31 | Google Llc | Audio-visual speech separation |
Non-Patent Citations (1)
Title |
---|
EPHRAT, A. et al., ‘Looking to Listen at the Cocktail Party: A Speaker-Independent Audio-Visual Model for Speech Separation’, 9 August 2018 [retrieved from the web on 7 February 2022] * |
Also Published As
Publication number | Publication date |
---|---|
CN113035227B (en) | 2022-02-11 |
CN113035227A (en) | 2021-06-25 |
AU2022200439A1 (en) | 2022-09-29 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
AU2022200439B2 (en) | Multi-modal speech separation method and system | |
Rivet et al. | Mixing audiovisual speech processing and blind source separation for the extraction of speech signals from convolutive mixtures | |
CN110600017A (en) | Training method of voice processing model, voice recognition method, system and device | |
Gogate et al. | DNN driven speaker independent audio-visual mask estimation for speech separation | |
Chen et al. | The first multimodal information based speech processing (misp) challenge: Data, tasks, baselines and results | |
CN108899044A (en) | Audio signal processing method and device | |
Pan et al. | Selective listening by synchronizing speech with lips | |
WO2022048239A1 (en) | Audio processing method and device | |
US20230335148A1 (en) | Speech Separation Method, Electronic Device, Chip, and Computer-Readable Storage Medium | |
Chen et al. | Multi-Modality Matters: A Performance Leap on VoxCeleb. | |
Yu et al. | Audio-visual multi-channel integration and recognition of overlapped speech | |
Wang et al. | A network model of speaker identification with new feature extraction methods and asymmetric BLSTM | |
CN116472579A (en) | Machine learning for microphone style transfer | |
Chen et al. | Sound localization by self-supervised time delay estimation | |
Li et al. | VCSE: Time-domain visual-contextual speaker extraction network | |
Ahmad et al. | Speech enhancement for multimodal speaker diarization system | |
Li et al. | Rethinking the visual cues in audio-visual speaker extraction | |
Wang et al. | The dku post-challenge audio-visual wake word spotting system for the 2021 misp challenge: Deep analysis | |
Pan et al. | ImagineNet: Target speaker extraction with intermittent visual cue through embedding inpainting | |
Suresh et al. | Computer-aided interpreter for hearing and speech impaired | |
Gul et al. | A survey of audio enhancement algorithms for music, speech, bioacoustics, biomedical, industrial and environmental sounds by image U-Net | |
Abel et al. | Cognitively inspired audiovisual speech filtering: towards an intelligent, fuzzy based, multimodal, two-stage speech enhancement system | |
CN115691539A (en) | Two-stage voice separation method and system based on visual guidance | |
Liu et al. | Multi-Modal Speech Separation Based on Two-Stage Feature Fusion | |
Zermini | Deep Learning for Speech Separation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FGA | Letters patent sealed or granted (standard patent) |