CN111929638A - Voice direction of arrival estimation method and device - Google Patents

Voice direction of arrival estimation method and device Download PDF

Info

Publication number
CN111929638A
CN111929638A CN202011011975.6A CN202011011975A CN111929638A CN 111929638 A CN111929638 A CN 111929638A CN 202011011975 A CN202011011975 A CN 202011011975A CN 111929638 A CN111929638 A CN 111929638A
Authority
CN
China
Prior art keywords
arrival
voice
grid
frequency
signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011011975.6A
Other languages
Chinese (zh)
Inventor
谭祚
何云鹏
许兵
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chipintelli Technology Co Ltd
Original Assignee
Chipintelli Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chipintelli Technology Co Ltd filed Critical Chipintelli Technology Co Ltd
Priority to CN202011011975.6A priority Critical patent/CN111929638A/en
Publication of CN111929638A publication Critical patent/CN111929638A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S3/00Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received
    • G01S3/80Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received using ultrasonic, sonic or infrasonic waves
    • G01S3/802Systems for determining direction or deviation from predetermined direction
    • G01S3/8027By vectorial composition of signals received by plural, differently-oriented transducers

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Circuit For Audible Band Transducer (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)

Abstract

A method for estimating a direction of arrival of speech includes the steps of S1: splitting a broadband voice signal received by a microphone array into a plurality of narrowband voice signals; dividing the space domain into a plurality of initial space domain grids S2: calculating a guide vector and a covariance matrix of each frequency point in each narrowband voice signal, separating a signal subspace and a noise subspace, and solving the space spectrum energy of each grid direction; judging the estimated direction of arrival through grid points corresponding to the spatial spectrum energy peak; s3: increasing grid points in the grid interval where the target may exist through the direction of arrival obtained in step S2; the step S2 is repeated on the reduced mesh after the mesh point is added to correct the direction of arrival until the mesh addition upper limit is reached. By adopting the voice direction of arrival estimation method, the mesh points are added in the estimation process in a self-adaptive manner; the estimation precision and the resolution of the voice signal direction of arrival method can be improved.

Description

Voice direction of arrival estimation method and device
Technical Field
The invention belongs to the technical field of intelligent voice recognition, relates to voice front-end signal processing, and particularly relates to a method and a device for estimating a voice direction of arrival.
Background
The direction-of-arrival estimation algorithm is mainly used for estimating the angle information of the sound source target and the microphone array, and inputting the estimated angle data into the voice enhancement system to effectively enhance the voice signals in the direction and inhibit the noise signals in other directions. Currently, the arrival direction estimation algorithm of the speech signal usually adopts a method based on the arrival time difference for estimation. However, when the signal-to-noise ratio of the input signal is lowered, reverberation is severe. The performance of such algorithms will drop significantly. Moreover, when the number of target sound sources increases, the method cannot distinguish the correct number of target sound sources, thereby causing performance failure.
The prior art has the following disadvantages:
1. the reverberation resistance is weak;
2. the noise immunity is weak;
3. the estimated resolution is low;
4. the estimation accuracy is insufficient.
Disclosure of Invention
In order to overcome the defects in the prior art, the invention discloses a method and a device for estimating the direction of arrival of voice.
The invention relates to a method for estimating the direction of arrival of voice, which comprises the following steps:
s1: splitting a broadband voice signal received by a microphone array into a plurality of narrowband voice signals; dividing a 0-180-degree airspace into a plurality of initial airspace grids;
s2: calculating a guide vector and a covariance matrix of each frequency point in each narrowband voice signal, separating a signal subspace and a noise subspace, and solving the space spectrum energy of each grid point of an airspace grid; judging the estimated direction of arrival through grid points corresponding to the spatial spectrum energy peak;
s3: increasing grid points in the grid interval where the target may exist through the direction of arrival obtained in step S2; the step S2 is repeated on the reduced mesh after the mesh point is added to correct the direction of arrival until the mesh addition upper limit is reached.
Preferably, the specific step of splitting the wideband speech signal into a plurality of narrowband speech signals in step S1 is: windowing and framing the voice signals, converting the time domain audio signals of each frame into a frequency domain, carrying out frequency division processing on the frequency spectrum of the frequency domain signals, and dividing the broadband voice signals into a plurality of narrow-band voice signals.
Preferably, in step S2, for each narrow band, the narrow band is the second in-band
Figure 469372DEST_PATH_IMAGE001
The steering vector of each frequency point is expressed as:
Figure 557151DEST_PATH_IMAGE002
wherein the content of the first and second substances,
Figure 274572DEST_PATH_IMAGE003
representing the grid point angles of a single initial spatial grid, the subscripts 1, 2 … k representing the different grid points,
Figure 607464DEST_PATH_IMAGE004
indicates the frequency of the ith frequency point,
Figure 93940DEST_PATH_IMAGE005
representing the distance between two microphones in a microphone array,
Figure 618200DEST_PATH_IMAGE006
is the speed of sound, e is a natural constant, and j represents the complex imaginary component.
Preferably, the covariance matrix R of the frequency points in step S2yyExpressed as:
Figure 822917DEST_PATH_IMAGE007
wherein
Figure 959500DEST_PATH_IMAGE008
Representing the frequency domain of the received audio signal,
Figure 566062DEST_PATH_IMAGE009
in the form of an array of flow pattern matrices,
Figure 1503DEST_PATH_IMAGE010
which represents the variance of the noise, is,
Figure 959095DEST_PATH_IMAGE011
is an identity matrix, RyyRepresenting received signals of a microphone arrayThe covariance matrix of (a) is determined,
Figure 633790DEST_PATH_IMAGE012
the covariance matrix of the voice signal is represented, the superscript H represents the conjugate transpose operation, theta is the direction of arrival of the voice signal, f is the frequency point frequency, and E is the expected value calculation symbol.
Further, the step S2 of separating the signal subspace and the noise subspace specifically includes:
covariance matrix R of received signals of microphone arrayyyDecomposed into signal subspace and noise subspace by eigenvalue decomposition method, expressed as:
Figure 94858DEST_PATH_IMAGE013
wherein
Figure 429762DEST_PATH_IMAGE014
A signal sub-space is represented that is,
Figure 609071DEST_PATH_IMAGE015
representing a noise subspace; sigmaX、ΣNRespectively representing a diagonal matrix of eigenvalues of the signal subspace and the noise subspace.
Further, in step S2, the spatial spectral energy in each grid direction is estimated using the orthogonality between the steering vector and the noise subspace, and the spatial spectral energy P (θ, f) in each grid direction is expressed as:
Figure 821877DEST_PATH_IMAGE016
wherein the content of the first and second substances,
Figure 635987DEST_PATH_IMAGE017
grid point corresponding to the peak value of
Figure 908837DEST_PATH_IMAGE003
I.e. the estimated direction of arrival,
Figure 575442DEST_PATH_IMAGE018
indicating a steering vector, superscript H indicating a conjugate transpose operation,
Figure 591939DEST_PATH_IMAGE015
representing a noise subspace, theta is the direction of arrival of the voice signal, and f is the frequency point frequency.
Preferably, in step S3, the grid points are symmetrically added on both sides of the grid point where the estimated direction of arrival is located, and the added grid points should be located in the last divided minimum grid.
The invention also discloses a voice direction-of-arrival estimation device, which comprises an array module, a control module, an input module, an estimation module, an optimization module and an output module which are sequentially connected;
the array module is a microphone array formed by a plurality of microphones arranged on a horizontal plane;
the control module realizes that the working state of the voice direction-of-arrival estimation algorithm is controlled by the identification of the awakening word;
the input module processes the voice signal and converts the voice signal into frequency point data with different frequencies;
the estimation module selects a frequency band range with obvious voice characteristics, and carries out primary direction-of-arrival estimation on each frequency point data in the range;
the optimization module optimizes an algorithm by a self-adaptive method for increasing spatial domain grid points;
and the output module transmits the estimated arrival direction of the voice to the voice enhancement system for enhancing the voice by a subsequent system.
By adopting the method and the device for estimating the direction of arrival of the voice, the grid points are added in an adaptive manner in the estimation process; the estimation precision and the resolution of the voice signal direction of arrival method can be improved.
Drawings
FIG. 1 is a schematic diagram of an embodiment of the estimation method according to the present invention;
FIG. 2 is a schematic view of an embodiment of the apparatus according to the present invention;
fig. 3 is a schematic diagram of an embodiment of the present invention, in which grid points are added, and the coordinate axes in fig. 3 represent angles.
Detailed Description
The following provides a more detailed description of the present invention.
The method for estimating the direction of arrival of voice, as shown in fig. 1, includes the following steps:
s1: splitting a broadband voice signal received by a microphone array into a plurality of narrowband voice signals; dividing a 0-180-degree airspace into a plurality of initial airspace grids;
s2: calculating a guide vector and a covariance matrix of each frequency point in each narrowband voice signal, separating a signal subspace and a noise subspace, and solving the space spectrum energy of each grid point of an airspace grid; judging the estimated direction of arrival through grid points corresponding to the spatial spectrum energy peak;
s3: increasing grid points in the grid interval where the target may exist through the direction of arrival obtained in step S2; the step S2 is repeated on the reduced mesh after the mesh point is added to correct the direction of arrival until the mesh addition upper limit is reached.
Specifically, the splitting the wideband speech signal into a plurality of narrowband speech signals in step S1 includes: windowing and framing the voice signals, converting the time domain audio signals of each frame into a frequency domain, carrying out frequency division processing on the frequency spectrum of the frequency domain signals, and dividing the broadband voice signals into a plurality of narrow-band voice signals;
one specific implementation way is that the time domain audio signal of each frame is converted into the frequency domain by fast fourier transform of 512 points, so as to obtain a plurality of discrete frequency points, for example, each frequency point in the frequency band interval of 1KHZ-3KHZ is subjected to frequency division processing, and under the sampling rate of 16KHZ, the broadband voice signal can be divided into 71 narrow bands, and each narrow band represents one discrete frequency point.
And dividing the 0-180-degree space domain into a plurality of initial space domain grids, and calculating the guide vector of each frequency point of the space domain grids.
For example, at 20 degree meshThe size of the grid is divided into 9 initial space domain grid intervals by 0-180 degrees to obtain 10 grid points, and then theta is calculated1Is 0 degree, theta2Is 20 degrees … theta10Is 180 degrees.
In step S2, a guide vector and a covariance matrix of each frequency point in each narrowband speech signal are calculated, a signal subspace and a noise subspace are separated by using the guide vector and the covariance matrix, and spatial spectrum energy in each grid direction is solved; the method specifically comprises the following steps:
s21 for each narrowband, the steering vector of the ith frequency point in the narrowband can be represented as:
Figure 293179DEST_PATH_IMAGE002
wherein the content of the first and second substances,
Figure 501044DEST_PATH_IMAGE003
representing the grid point angles of a single initial spatial grid, the subscripts 1, 2 … k representing the different grid points,
Figure 654945DEST_PATH_IMAGE004
indicates the frequency of the ith frequency point,
Figure 475133DEST_PATH_IMAGE005
representing the distance between two microphones in a microphone array,
Figure 30880DEST_PATH_IMAGE006
for the speed of sound, E is a natural constant, j represents the complex imaginary component, and E is the expected value calculation sign.
S22 estimating covariance matrix R of each frequency point of input signalYYIt can be expressed as:
Figure 427224DEST_PATH_IMAGE007
wherein
Figure 68421DEST_PATH_IMAGE008
Representing the frequency domain of the received audio signal,
Figure 692301DEST_PATH_IMAGE009
in the form of an array of flow pattern matrices,
Figure 305816DEST_PATH_IMAGE010
which represents the variance of the noise, is,
Figure 855483DEST_PATH_IMAGE011
is an identity matrix, RyyA covariance matrix representing the received signals of the microphone array,
Figure 718397DEST_PATH_IMAGE012
the covariance matrix of the voice signal is represented, the superscript H represents the conjugate transpose operation, theta is the direction of arrival of the voice signal, and f is the frequency point frequency.
Decomposing the estimated covariance matrix into a signal subspace and a noise subspace by an eigenvalue decomposition method can be expressed as:
Figure 145967DEST_PATH_IMAGE013
,
wherein
Figure 879568DEST_PATH_IMAGE014
A signal sub-space is represented that is,
Figure 865717DEST_PATH_IMAGE015
representing a noise subspace; sigmaX、ΣNRespectively representing a diagonal matrix of eigenvalues of the signal subspace and the noise subspace.
And finally, estimating the spatial spectrum energy of each grid direction by using the orthogonal characteristics of the guide vector and the noise subspace, wherein the spatial spectrum energy P (theta, f) of each grid direction can be expressed as:
Figure 950347DEST_PATH_IMAGE016
wherein the content of the first and second substances,
Figure 916029DEST_PATH_IMAGE017
grid point corresponding to the peak value of
Figure 300874DEST_PATH_IMAGE003
I.e. the estimated direction of arrival,
Figure 926765DEST_PATH_IMAGE018
indicating a steering vector, superscript H indicating a conjugate transpose operation,
Figure 29851DEST_PATH_IMAGE015
representing a noise subspace, theta is the direction of arrival of the voice signal, and f is the frequency point frequency.
After the first estimated direction of arrival is obtained according to the first divided mesh, the range of the direction of arrival is narrowed by continuously adding the mesh points because the range of the first divided mesh is large and the real direction of arrival is near the first estimated direction of arrival.
S3 symmetrically adds grid points on both sides of the grid point where the estimated direction of arrival is located, where the added grid points should be located in the minimum grid of the last division, for example, the distance between the new grid point and the center grid point is usually a reduction ratio of the last grid width.
After the grid points are added, the steps S21-S22 are continuously repeated in the interval of the added pair of grid points, and the grid point where the updated estimated direction of arrival is located is obtained;
step S3 is repeated until the grid point addition number reaches the grid addition upper limit, which is usually set in consideration of the hardware condition, the time required for calculation, and the like.
As shown in fig. 3, if the first estimation estimates the direction of arrival at the 80 degree grid points, the first incremental grid points W1 may be added at 70 degrees and 90 degrees to the left and right of the 80 degree grid points, steps S21-S22 may be repeated to find a new estimated direction of arrival at the 70 degree grid points, the second incremental grid points W2 may be added at 65 and 75 degrees to the left and right of the 70 degree grid points, and the new direction of arrival may be estimated again at 75 degrees, and the pair of third incremental grid points W3 may continue to be added at the left and right of 75 degrees. And in the same way, continuously adding grid points and reducing the grid point gap until the added grid points reach the upper limit.
The spatial distance of the spatial grid directly affects the estimation accuracy and resolution of the algorithm, but the densely distributed spatial grid increases the computational power of the algorithm and is not beneficial to engineering application. The invention adopts the airspace grids with larger space intervals at the beginning, counts the estimated prior arrival direction after the initial estimation of the first frequency points, and then adds new grid points in the grid interval of the prior direction to increase the number of the airspace grids continuously so as to present non-uniform distribution, so that in the airspace grid interval where the target possibly exists, the grids become dense and the grids in other intervals are sparse, and the estimation precision and the resolution of the estimation of the arrival direction can be improved.
The estimation method of the present invention can be implemented based on a device, as shown in fig. 2, which includes an array module, a control module, an input module, an estimation module, an optimization module, and an output module, which are connected in sequence. The functions of each module are as follows:
the array module is provided with a microphone array formed by a plurality of microphones on a horizontal plane to receive voice signals. The array structure, such as a circular array, a linear array and the like, can be set according to application requirements;
the control module realizes that the working state of the voice direction-of-arrival estimation algorithm is controlled by the recognition of the awakening word, and when the awakening word is recognized, the system starts to operate the voice direction-of-arrival estimation algorithm;
the input module can firstly carry out windowing and framing processing on the voice signals and convert non-stationary voice signals into short-time stationary signals. Then, converting the voice signal of each frame into frequency point data with different frequencies through a fast Fourier transform algorithm, and estimating the direction of arrival of the voice signal through the frequency point data;
and the estimation module selects a frequency band range with obvious voice characteristics and carries out primary direction-of-arrival estimation on each frequency point data in the range. The signal space of each frequency point can be decomposed into a signal subspace and a noise subspace through a characteristic value decomposition algorithm. Then, setting spatial grids with the same spatial interval, calculating the spatial spectrum energy of each spatial grid point by using the orthogonal characteristic of the guide vector and the noise subspace, and determining the arrival direction of the voice signal by retrieving the peak value of the spatial spectrum energy;
the optimization module adopts spatial grids with large spatial intervals when estimation is started, counts the estimated prior direction of arrival after initial estimation of a plurality of previous frequency points, and adds new grid points in the grid interval of the prior direction, so that the number of the spatial grids is continuously increased to present non-uniform distribution. The optimization module optimizes the performance of the algorithm by a method of increasing spatial domain grid points in a self-adaptive mode.
And the output module is used for transmitting the estimated speech arrival direction to the speech enhancement system for enhancing the speech by a subsequent system.
The foregoing is directed to preferred embodiments of the present invention, wherein the preferred embodiments are not obviously contradictory or subject to any particular embodiment, and any combination of the preferred embodiments may be combined in any overlapping manner, and the specific parameters in the embodiments and examples are only for the purpose of clearly illustrating the inventor's invention verification process and are not intended to limit the scope of the invention, which is defined by the claims and the equivalent structural changes made by the description and drawings of the present invention are also intended to be included in the scope of the present invention.

Claims (8)

1. A method for estimating the direction of arrival of speech, comprising the steps of:
s1: splitting a broadband voice signal received by a microphone array into a plurality of narrowband voice signals; dividing a 0-180-degree airspace into a plurality of initial airspace grids;
s2: calculating a guide vector and a covariance matrix of each frequency point in each narrowband voice signal, separating a signal subspace and a noise subspace, and solving the space spectrum energy of each grid point of an airspace grid; judging the estimated direction of arrival through grid points corresponding to the spatial spectrum energy peak;
s3: increasing grid points in the grid interval where the target may exist through the direction of arrival obtained in step S2; the step S2 is repeated on the reduced mesh after the mesh point is added to correct the direction of arrival until the mesh addition upper limit is reached.
2. The method for estimating the direction of arrival of speech according to claim 1, wherein the step S1 of splitting the wideband speech signal into a plurality of narrowband speech signals comprises the following steps: windowing and framing the voice signals, converting the time domain audio signals of each frame into a frequency domain, carrying out frequency division processing on the frequency spectrum of the frequency domain signals, and dividing the broadband voice signals into a plurality of narrow-band voice signals.
3. The method for estimating a direction of arrival of a speech according to claim 1, wherein in said step S2, for each narrow band, the number of narrow bands is the second
Figure 164809DEST_PATH_IMAGE001
The steering vector of each frequency point is expressed as:
Figure 295838DEST_PATH_IMAGE002
wherein the content of the first and second substances,
Figure 640101DEST_PATH_IMAGE003
representing the grid point angles of a single initial spatial grid, the subscripts 1, 2 … k representing the different grid points,
Figure 964903DEST_PATH_IMAGE004
indicates the frequency of the ith frequency point,
Figure 161135DEST_PATH_IMAGE005
representing the distance between two microphones in a microphone array,
Figure 238682DEST_PATH_IMAGE006
for the speed of sound, E is a natural constant, j represents the complex imaginary component, and E is the expected value calculation sign.
4. The method according to claim 1, wherein the covariance matrix R of the frequency points in step S2 isyyExpressed as:
Figure 973420DEST_PATH_IMAGE007
wherein
Figure 270671DEST_PATH_IMAGE008
Representing the frequency domain of the received audio signal,
Figure 240901DEST_PATH_IMAGE009
in the form of an array of flow pattern matrices,
Figure 923686DEST_PATH_IMAGE010
which represents the variance of the noise, is,
Figure 78593DEST_PATH_IMAGE011
is an identity matrix, RyyA covariance matrix representing the signals received by the microphone array,
Figure 846829DEST_PATH_IMAGE012
the covariance matrix of the voice signal is represented, the superscript H represents the conjugate transpose operation, theta is the direction of arrival of the voice signal, f is the frequency point frequency, and E is the expected value calculation symbol.
5. The method according to claim 4, wherein the step of separating the signal subspace and the noise subspace in step S2 is specifically as follows:
covariance matrix R of received signals of microphone arrayyyDecomposed into signal subspace and noise subspace by eigenvalue decomposition method, expressed as:
Figure 261497DEST_PATH_IMAGE013
wherein
Figure 798788DEST_PATH_IMAGE014
A signal sub-space is represented that is,
Figure 124596DEST_PATH_IMAGE015
representing a noise subspace; sigmaX、ΣNRespectively representing a diagonal matrix of eigenvalues of the signal subspace and the noise subspace.
6. The speech direction-of-arrival estimation method according to claim 4, wherein in step S2, the spatial spectral energy of each grid direction is estimated using the orthogonality between the steering vector and the noise subspace, and the spatial spectral energy P (θ, f) of each grid direction is represented as:
Figure 504762DEST_PATH_IMAGE016
wherein the content of the first and second substances,
Figure 692161DEST_PATH_IMAGE017
grid point corresponding to the peak value of
Figure 100271DEST_PATH_IMAGE003
I.e. the estimated direction of arrival,
Figure 347713DEST_PATH_IMAGE018
indicating a steering vector, superscript H indicating a conjugate transpose operation,
Figure 605388DEST_PATH_IMAGE015
representing a noise subspace, theta is the direction of arrival of the voice signal, and f is the frequency point frequency.
7. The method for estimating a speech direction of arrival according to claim 1, wherein in step S3, the addition points are symmetrically added on both sides of the point where the estimated direction of arrival is located, and the addition points should be located within the minimum mesh of the last division.
8. A voice direction-of-arrival estimation device is characterized by comprising an array module, a control module, an input module, an estimation module, an optimization module and an output module which are sequentially connected;
the array module is a microphone array formed by a plurality of microphones arranged on a horizontal plane;
the control module realizes that the working state of the voice direction-of-arrival estimation algorithm is controlled by the identification of the awakening word;
the input module processes the voice signal and converts the voice signal into frequency point data with different frequencies;
the estimation module selects a frequency band range with obvious voice characteristics, and carries out primary direction-of-arrival estimation on each frequency point data in the range;
the optimization module optimizes an algorithm by a self-adaptive method for increasing spatial domain grid points;
and the output module transmits the estimated arrival direction of the voice to the voice enhancement system for enhancing the voice by a subsequent system.
CN202011011975.6A 2020-09-24 2020-09-24 Voice direction of arrival estimation method and device Pending CN111929638A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011011975.6A CN111929638A (en) 2020-09-24 2020-09-24 Voice direction of arrival estimation method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011011975.6A CN111929638A (en) 2020-09-24 2020-09-24 Voice direction of arrival estimation method and device

Publications (1)

Publication Number Publication Date
CN111929638A true CN111929638A (en) 2020-11-13

Family

ID=73335026

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011011975.6A Pending CN111929638A (en) 2020-09-24 2020-09-24 Voice direction of arrival estimation method and device

Country Status (1)

Country Link
CN (1) CN111929638A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112485761A (en) * 2021-02-03 2021-03-12 成都启英泰伦科技有限公司 Sound source positioning method based on double microphones
CN113203987A (en) * 2021-07-05 2021-08-03 成都启英泰伦科技有限公司 Multi-sound-source direction estimation method based on K-means clustering

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101426862B1 (en) * 2013-03-19 2014-08-07 국방과학연구소 3 Dimension Array Antenna System and Altitude Angle Estimation Method thereof
CN108802667A (en) * 2018-05-25 2018-11-13 哈尔滨工程大学 Wave arrival direction estimating method based on generalized orthogonal match tracing
CN108872926A (en) * 2018-07-11 2018-11-23 哈尔滨工程大学 A kind of amplitude and phase error correction and DOA estimation method based on convex optimization
CN109407045A (en) * 2018-10-10 2019-03-01 苏州大学 A kind of non-homogeneous sensor array broadband signal Wave arrival direction estimating method
CN109783960A (en) * 2019-01-23 2019-05-21 桂林电子科技大学 A kind of Wave arrival direction estimating method based on meshing refinement

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR101426862B1 (en) * 2013-03-19 2014-08-07 국방과학연구소 3 Dimension Array Antenna System and Altitude Angle Estimation Method thereof
CN108802667A (en) * 2018-05-25 2018-11-13 哈尔滨工程大学 Wave arrival direction estimating method based on generalized orthogonal match tracing
CN108872926A (en) * 2018-07-11 2018-11-23 哈尔滨工程大学 A kind of amplitude and phase error correction and DOA estimation method based on convex optimization
CN109407045A (en) * 2018-10-10 2019-03-01 苏州大学 A kind of non-homogeneous sensor array broadband signal Wave arrival direction estimating method
CN109783960A (en) * 2019-01-23 2019-05-21 桂林电子科技大学 A kind of Wave arrival direction estimating method based on meshing refinement

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
朱鹏等: ""基于网格-粒子群的二维DOA 估计"", 《成都工业学院学报》 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112485761A (en) * 2021-02-03 2021-03-12 成都启英泰伦科技有限公司 Sound source positioning method based on double microphones
CN113203987A (en) * 2021-07-05 2021-08-03 成都启英泰伦科技有限公司 Multi-sound-source direction estimation method based on K-means clustering

Similar Documents

Publication Publication Date Title
CN109830245B (en) Multi-speaker voice separation method and system based on beam forming
US10901063B2 (en) Localization algorithm for sound sources with known statistics
CN108986838B (en) Self-adaptive voice separation method based on sound source positioning
CN109272989B (en) Voice wake-up method, apparatus and computer readable storage medium
Araki et al. Underdetermined blind sparse source separation for arbitrarily arranged multiple sensors
CN109490822B (en) Voice DOA estimation method based on ResNet
US10127922B2 (en) Sound source identification apparatus and sound source identification method
CN110223708B (en) Speech enhancement method based on speech processing and related equipment
CN104076331A (en) Sound source positioning method for seven-element microphone array
US8521477B2 (en) Method for separating blind signal and apparatus for performing the same
CN103854660B (en) A kind of four Mike's sound enhancement methods based on independent component analysis
CN108091345B (en) Double-ear voice separation method based on support vector machine
Saruwatari et al. Blind source separation for speech based on fast-convergence algorithm with ICA and beamforming
CN111929638A (en) Voice direction of arrival estimation method and device
CN112034418A (en) Beam scanning method based on frequency domain Bark sub-band and sound source orientation device
CN108538306B (en) Method and device for improving DOA estimation of voice equipment
CN107219512A (en) A kind of sound localization method based on acoustic transfer function
CN110709929A (en) Processing sound data to separate sound sources in a multi-channel signal
CN114245266A (en) Area pickup method and system for small microphone array device
Yin et al. Multi-talker Speech Separation Based on Permutation Invariant Training and Beamforming.
CN116559778B (en) Vehicle whistle positioning method and system based on deep learning
Gao et al. A modified frequency weighted MUSIC algorithm for multiple sound sources localization
CN111060867A (en) Directional microphone microarray direction of arrival estimation method
CN114639398B (en) Broadband DOA estimation method based on microphone array
CN116153324A (en) Virtual array expansion beam forming method based on deep learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20201113

RJ01 Rejection of invention patent application after publication