CN103995252A - Three-dimensional space sound source positioning method - Google Patents

Three-dimensional space sound source positioning method Download PDF

Info

Publication number
CN103995252A
CN103995252A CN201410202062.0A CN201410202062A CN103995252A CN 103995252 A CN103995252 A CN 103995252A CN 201410202062 A CN201410202062 A CN 201410202062A CN 103995252 A CN103995252 A CN 103995252A
Authority
CN
China
Prior art keywords
microphone
tau
sound
formula
overbar
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201410202062.0A
Other languages
Chinese (zh)
Other versions
CN103995252B (en
Inventor
郭业才
朱赛男
张宁
黄友锐
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN201410202062.0A priority Critical patent/CN103995252B/en
Publication of CN103995252A publication Critical patent/CN103995252A/en
Application granted granted Critical
Publication of CN103995252B publication Critical patent/CN103995252B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S5/00Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
    • G01S5/18Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using ultrasonic, sonic, or infrasonic waves
    • G01S5/22Position of source determined by co-ordinating a plurality of position lines defined by path-difference measurements

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

The invention provides a three-dimensional space sound source positioning method, and belongs to the technical field of sound source positioning. The method includes the steps that a double-L-shaped microphone array is established; on the basis of the normalization frequency domain minimum mean square method, a penalty function is introduced to correct frequency spectrum energy, and then impulse responses of different array elements are adaptively estimated and delay inequality is calculated; position coordinates of a sound source are determined through the relation between the acquired delay inequality and the position of double-L-shaped microphones. Compared with a traditional positioning method, the array structure is simple, the calculated quantity is small, anti-noise performance, reverberation resistance and positioning accuracy are effectively improved, and the method is more suitable for indoor three-dimensional sound source positioning and can be widely applied to a car hands-free phone, a video conference system, a speech recognition system, an intelligent robot and other fields.

Description

A kind of three dimensions sound localization method
Technical field
The invention belongs to auditory localization technical field, especially relate to a kind of method of utilizing microphone array to carry out sound source three-dimensional location.
Background technology
At present, auditory localization based on microphone array is a major issue in acoustics signal processing field, compared to traditional Array Signal Processing, the voice signal of array microphone processing does not have carrier wave, the range of signal that can process is wide, adaptable, be all widely used in fields such as vehicle-carried hands-free telephone, video conferencing system, speech recognition system and intelligent robots.Auditory localization based on microphone array mainly contain based on steerable beam form localization method, based on arrive delay inequality localization method and based on high-resolution localization method.Wherein, the voice signal that the localization method forming based on steerable beam receives microphone array carries out filtering, weighted sum, then directly control microphone and point to the direction that makes wave beam have peak power output, because its computation complexity height can not be used to real time processing system; The correlation matrix solving between Mike's signal based on high-resolution localization method utilization is made deflection, thereby further make sound source position, although be successfully applied to the processing of some array signals, but locating effect is not good in actual applications, be not only subject to array structure restriction and also signal not steadily time calculated amount can be multiplied; And wherein, be by estimating that sound source is to the delay inequality between microphone and according to estimating sound source position in the position of microphone based on arriving the sound localization method of delay inequality, be not subject to array structure restriction, calculated amount little.
Sound localization method based on arriving delay inequality mainly divides two parts to complete: time delay is estimated and location.Delay time estimation method mainly contains broad sense cross-correlation method (Generalized Cross Correlation, GCC), minimum mean square error method (Least Mean Square, and self-adaptive features value decomposition method (Adaptive Engenvalue Decomposition, AED) LMS).Wherein, GCC method performance under reverberation environment can decline a lot; LMS method performance is basic suitable with it; AED method is by estimating that double-channel impulse response suppresses reverberation, but it requires double-channel relatively prime, does not contain common zero point.But in actual indoor environment, impulse response length is generally very long, the relatively prime possibility of double-channel is very little, and AED method is no longer applicable.In order to improve relatively prime possibility, AED binary channels method of estimation is generalized to self-adapting multi-channel method of estimation (Adaptive Multichannel by the people such as Y (Arden) Huang, AMC), proposition utilizes normalization hyperchannel frequency domain least-square methods (Normalized Multichannel Frequency-domain Least Mean Square, NMCFLMS) to estimate the impulse response of each array element.The basic thought of NMCFLMS method is, observation signal is divided into continuous block signal, adopt frequency domain normalization minimum mean-square method to carry out channel estimation in frequency domain, though there is the feature of the low and Newton method fast convergence rate of hyperchannel frequency domain least-square methods complexity, but in the time that channel additive noise exists, may disperse, can not effectively carry out channel estimating.
Localization method mainly contains least square method and geometry location method, the former calculation of complex, and to initial value sensitivity; The latter utilizes microphone position parameter and time delay estimated value to determine bi-curved funtcional relationship, and multiple bi-curved intersection points are definite sound source position, calculates simply, but easily occurs not exclusive closed solution, is difficult to effective location.
Summary of the invention
May disperse in the time that channel additive noise exists for sound localization method in prior art, can not effectively carry out channel estimating and be difficult to the defect of effective location sound source, the invention provides a kind of improved three dimensions sound localization method, utilize the method for revising normalization hyperchannel frequency domain least-square methods (MNMCFLMS) and limiting channel impulse response length to carry out channel estimating, effectively improved the precision of auditory localization.
In order to achieve the above object, the invention provides following technical scheme:
A kind of three dimensions sound localization method, comprises the steps:
Steps A, in three-dimensional cartesian coordinate system, the L-type microphone array of setting up two at grade and being oppositely arranged;
Step B, adopts the normalization hyperchannel frequency domain least-square methods of revising to estimate that sound-source signal arrives the delay inequality of each microphone, comprises the steps:
Step B-1, sound-source signal s (n) is through the channel impulse response h of i microphone i(n) rear and channel additive noise v i(n) merge, obtain the reception signal x of i microphone i(n): x i(n)=s t(n) h i(n)+v i(n);
Wherein, T representing matrix matrix transpose operation; N is time series, is integer;
In the time disregarding channel additive noise, the reception signal x of i microphone i(n) with the reception signal x of j microphone j(n) pass between is:
x i T ( n ) h j ( n ) = x j T ( n ) h i ( n ) ;
Step B-2, with length be 2L hthe reception signal x of rectangular window function w (n) to i microphone i(n) carry out windowing process, the n frame that obtains i microphone receives signal and is:
x i ( n ) 2 L h × 1 = [ x i ( n L h - L h ) , x i ( n L h - L h + 1 ) , . . . , x i ( n L h + L h - 1 ) ] T ;
In formula, L hrepresent channel impulse response h i(n) length is integer; N=1 ..., int (N/L h-1); N is the sequence total length that microphone receives signal, is integer; Int (N/L h-1) be to N/L h-1 integer obtaining after rounding downwards;
Step B-3, utilizes Fourier transform that the time-domain signal obtaining in step B-2 is transformed to frequency-region signal:
X i ( n ) 2 L h × 1 = F 2 L h × 2 L h x i ( n ) 2 L h × 1 ;
In formula, for 2L h× 2L hdimension Fourier transform matrix;
Step B-4, introduces penalty to spectrum energy correction, estimates adaptively channel frequency domain response
As channel additive noise v i(n), while existence, error of frequency domain function definition is
e ‾ ij ( n ) = W L h × 2 L h 01 D i ( n ) W 2 L h × L h 10 h ^ ‾ j ( n ) - W L h × 2 L h 01 D j ( n ) W 2 L h × L h 10 h ^ ‾ i ( n )
In formula
D i = diag ( F 2 L h × 2 L h ( x i ( n ) 2 L h × 1 ) ) ;
W L h × 2 L h 01 = F L h × L h 0 L h × L h I L h × L h F 2 L h × 2 L h - 1 ;
W 2 L h × L h 10 = F 2 L h × 2 L h I L h × L h 0 L h × L h T F L h × L h - 1 ;
In formula, i ≠ j; I, j=1,2 ..., M, be respectively L h× L hdimension Fourier transform matrix and Fourier transform inverse matrix; Diag represents diagonal matrix; represent dimension unit matrix; represent L h× L hdimension null matrix;
Utilize penalty J p(n) by the cost function J of Lagrange multiplier β (n) to NMCFLMS f(n) revise, the cost function that obtains revising is
J mod(n)=J f(n)+β(n){-J p(n)}
In formula
β ( n ) = | ▿ J p H ( n ) ▿ J f ( n ) | | | ▿ J p ( n ) | | 2
In formula, the cost function gradient that the value of Lagrange multiplier β (n) is revised during by stable state equals zero, i.e. ▽ J f(n)=β (n) ▽ J p(n) time, obtain, H represents conjugate transpose;
By the cost function J revising mod(n), obtain channel frequency domain response more new formula be
h ^ ‾ i 10 ( n + 1 ) = h ^ ‾ i 10 ( n ) - ▿ J f 01 ( n ) + β ^ ( n ) ▿ J p 10 ( n )
Wherein,
▿ J p 01 ( n ) = F 2 L h × 2 L h 0 L h × L h I L h × L h T K ( n ) F 2 L h × 2 L h 0 L h × L h I L h × L h T h ^ ‾ ( n )
β ^ ( n ) = | ▿ J p 10 ( n ) H ▿ J f 01 ( n ) | | | ▿ J p 10 | | 2
▿ J f 01 ( n ) = μ ( P i ( n ) + δ I 2 L h × L h ) - 1 × Σ k = 1 M D k * ( n ) e ‾ ki 01 ( n )
In above formula,
P i ( n ) = λ P i ( n - 1 ) + ( 1 - λ ) × Σ k = 1 , k ≠ i M D k * ( n ) D k ( n )
e ‾ ki 01 ( n ) = F 2 L h × 2 L h 0 L h × L h I L h × L h T F L h × L h - 1 e ‾ ki ( n )
In formula, K (n) is that diagonal entry is diagonal matrix; P i(n) be the spectrum energy of multi-channel output signal, can obtain more stable spectrum energy by forgetting factor λ, parameter δ is normal number, and it can effectively solve the noise scale-up problem that spectrum energy hour causes;
Step B-5, to channel frequency domain response carry out Fourier inversion, obtain channel impulse response estimation and be
h ^ i ( n ) = I L h × L h 0 L h × L h F 2 L h × 2 L h - 1 h ^ ‾ i 10 ( n )
In formula, for 2L h× 2L hdimension Fourier transform inverse matrix;
Step B-6, in adaptive process, the channel impulse response of estimation the corresponding time delay of the peak value of middle appearance is exactly the reception direct sound wave signal time delay of i microphone, and sound-source signal arrives the delay inequality between i microphone and j microphone and is
τ ^ ij = ( max l = 1 L h | h ^ il ( n ) | - max l = 1 L h | h ^ jl ( n ) | ) / f s
In formula, f sfor the sample frequency of signal, max represents to get maximal value;
Step C, is multiplied each other and is obtained sound-source signal and arrive the range difference of each microphone by the delay inequality described in step B and the velocity of sound, and according to the position relationship of each microphone, determine the position of described sound source.
Further, described J p(n) be the penalty under constraint condition, constraint condition is
max J P ( n ) = Σ i = 1 M Σ j = 0 L - 1 ln ( | h ^ ‾ ij | 2 ) s . t . Σ i = 1 M Σ j = 0 L h - 1 | h ^ ‾ ij | 2 = 1 ML h
In formula, at penalty J p(n) while maximization, constraint condition set up; S.t. represent constraint condition.
Further, the length L of channel impulse response in described step B-2 hrestrictive condition is:
L h ≤ 2 f s max i , j = 1 M { τ ij , max }
In formula, τ ij, max=d ij/ c is that between i and j microphone, maximum delay is poor, d ijbe the distance between i and j microphone, the velocity of propagation that c is sound source.
Further, in described steps A two L-type arrays to lay rule as follows: first microphone mic1 is placed in X-axis, and its coordinate is (d, 0,0); Second microphone mic2 and true origin are overlapping, and its coordinate is (0,0,0); The 3rd microphone mic3 is placed in Y-axis, and its coordinate is (0, d, 0), and d is distance between two adjacent microphones, is real number, and first, second and third microphone has formed first L-type microphone array; Meanwhile, taking x=D/2 as axis of symmetry, set up second the L-type microphone array relative with first L-type microphone array, wherein, the coordinate of the 4th microphone mic4 is (D, 0,0), the coordinate of the 5th microphone mic5 is (D-d, 0,0), the coordinate of the 6th microphone mic6 is (D, d, 0); D is two distances between relative L-type microphone array.
Further, in described step C, determine that the step of sound source position is as follows:
Step C-1, position is the delay inequality that the sound-source signal of (x, y, z) arrives i microphone and j microphone with i microphone and j microphone space from d ijpass be relation has following two groups in detail:
First group: x 2 + y 2 + z 2 = 1 2 c τ ^ 21 ( c 2 τ ^ 21 2 + 2 dx - d 2 ) With
x 2 + y 2 + z 2 = 1 2 c τ ^ 23 ( c 2 τ ^ 23 2 + 2 dy - d 2 ) ;
Second group: ( x - D ) 2 + y 2 + z 2 = 1 2 c τ ^ 45 ( c 2 τ ^ 45 2 d 2 - 2 d ( x - D ) ) With
( x - D ) 2 + y 2 + z 2 = 1 2 c τ ^ 46 ( c 2 τ ^ 46 2 + 2 Dx + D 2 - 2 dy - d 2 ) ;
In formula, for the delay inequality between sound-source signal to the second microphone and first microphone; for the delay inequality of sound-source signal to the second microphone and the 3rd microphone, for the delay inequality of four microphones of sound-source signal to the and the 5th microphone, for the delay inequality of four microphones of sound-source signal to the and the 6th microphone;
Step C-2, according to first group described in step C-1 with second group of relation, obtain the projection coordinate of sound source in xoy plane and be
x ^ = b 2 - b 1 a 1 - a 2
y ^ = a 1 b 2 - a 2 b 1 a 1 - a 2
z ^ = 0
In formula, a 1 = τ ^ 23 / τ ^ 21 ; a 2 = - τ ^ 46 / τ ^ 45 ; b 1 = ( c 2 τ ^ 21 τ ^ 23 + d 2 ) ( τ ^ 21 - τ ^ 23 ) / ( 2 d τ ^ 21 ) ; b 2 = ( ( c 2 τ ^ 45 τ ^ 46 + d 2 ) ( τ ^ 45 - τ ^ 46 ) + 2 dD τ ^ 45 ) / ( 2 d τ ^ 45 ) ;
Step C-3, projection coordinate by the sound source described in step C-2 in xoy plane in first group described in substitution step C-1 and four relational expressions of second group, obtains four z axial coordinate estimated values of sound source respectively getting the mean value of these four z coordinates estimates as the z coordinate of sound source for:
z ^ = z ^ 1 + z ^ 2 + z ^ 3 + z ^ 4 4 .
Compared with prior art, tool of the present invention has the following advantages and beneficial effect:
The present invention, on normalization frequency domain least-square methods basis, by introducing penalty, spectrum energy is revised, thereby it is poor to estimate adaptively impulse response the calculation delay of different array elements, avoids channel estimating to occur worsening; Utilize interior two the unique definite sound sources of crossing straight line of plane projected position planar, avoided the appearance of not exclusive closed solution problem; Define in addition channel impulse response length, strengthened the relatively prime property of double-channel, improved anti-reverberation ability.Compared with traditional localization method, array structure of the present invention is simple, calculated amount is few, noise robustness, anti-reverberation ability and positioning precision are effectively improved, more be suitable for indoor three-dimensional auditory localization, can be widely used in the every field such as vehicle-carried hands-free telephone, video conferencing system, speech recognition system and intelligent robot.
Brief description of the drawings
Fig. 1 is the flow chart of steps of three dimensions sound localization method provided by the invention;
Fig. 2 is double L-shaped microphone array setting principle figure;
Fig. 3 is the normalization hyperchannel frequency domain least-square methods schematic diagram of revising;
Fig. 4 is that in embodiment, microphone array is listed in the schematic diagram of putting in room;
Fig. 5 is normalized projection error convergence curve figure.
Embodiment
Below with reference to specific embodiment, technical scheme provided by the invention is elaborated, should understands following embodiment and only be not used in and limit the scope of the invention for the present invention is described.
First the present invention improves normalization hyperchannel frequency domain least-square methods NMCFLMS, increases its performance; Limit channel impulse response length simultaneously, strengthen the relatively prime property of double-channel; And utilize the NMCFLMS revising to carry out channel estimating.Specifically, process flow diagram as shown in Figure 1, comprises the steps:
Steps A, sets up three-dimensional coordinate system, at grade, 6 microphones is put into two relative L-type arrays, and between them, distance is D, and as shown in Figure 2: first microphone mic1 is placed in X-axis, its coordinate is (d, 0,0); Second microphone mic2 and true origin are overlapping, and its coordinate is (0,0,0); The 3rd microphone mic3 is placed in Y-axis, and its coordinate is (0, d, 0), and d is distance between two adjacent microphones, is real number, and first, second and third microphone has formed first L-type microphone array.Meanwhile, taking x=D/2 as axis of symmetry, set up second the L-type microphone array relative with first L-type microphone array, wherein, the coordinate of the 4th microphone mic4 is (D, 0,0), the coordinate of the 5th microphone mic5 is (D-d, 0,0), the coordinate of the 6th microphone mic6 is (D, d, 0); D is two distances between relative L-type microphone array.The projection coordinate of the coordinate (x, y, z) of sound source in xoy plane is (x, y, 0), and the distance between projection coordinate's point and initial point is γ, and both lines and x axle forward angle are θ.
Step B, adopts the normalization hyperchannel frequency domain least-square methods MNMCFLMS revising, and estimates that sound-source signal arrives the delay inequality of each microphone, its schematic diagram as shown in Figure 3:
Step B-1, the channel impulse response h of sound-source signal s (n), an i microphone i(n), channel additive noise v iand the reception signal x of i microphone (n) i(n) pass between is
x i(n)=s T(n)h i(n)+v i(n) (1)
In formula, T representing matrix matrix transpose operation; N is time series, is integer.
In the time disregarding channel additive noise, the reception signal x of i microphone i(n) with the reception signal x of j microphone j(n) pass between is
x i T ( n ) h j ( n ) = x j T ( n ) h i ( n ) - - - ( 2 )
In formula, i, j=1,2 ..., M, the number that M is microphone, is positive integer;
Step B-2, carries out frame to the reception signal of i microphone and moves as L hwindowing process: with length be 2L hrectangular window function w (n) is multiplied by the reception signal of i microphone, and the n frame that obtains i microphone receives signal and is:
x i ( n ) 2 L h × 1 = [ x i ( n L h - L h ) , x i ( n L h - L h + 1 ) , . . . , x i ( n L h + L h - 1 ) ] T - - - ( 3 )
In formula, L hrepresent channel impulse response h ilength, be integer; N=1 ..., int (N/L h-1); N is the sequence total length that microphone receives signal, is integer; Int (N/L h-1) be to N/L h-1 integer obtaining after rounding downwards;
In actual indoor environment, in order to overcome the very long very little deficiency of the relatively prime possibility of double-channel that causes of channel impulse response length, guaranteeing that double-channel is relatively prime when stronger, can be further to channel impulse response length L hlimit, restrictive condition is:
L h ≤ 2 f s max i , j = 1 M { τ ij , max } - - - ( 4 )
In formula, be that between i and j microphone, maximum delay is poor.By limiting channel impulse response length, effectively strengthen the relatively prime property of double-channel.
Step B-3, obtains frequency-region signal to formula (3) as Fourier transform
X i ( n ) 2 L h × 1 = F 2 L h × 2 L h x i ( n ) 2 L h × 1 - - - ( 5 )
In formula, for 2L h× 2L hdimension Fourier transform matrix.
Step B-4, introduces penalty to spectrum energy correction, estimates adaptively and estimates adaptively channel frequency domain response
As channel additive noise v i(n), while existence, error of frequency domain function definition is
e ‾ ij ( n ) = W L h × 2 L h 01 D i ( n ) W 2 L h × L h 10 h ^ ‾ j ( n ) - W L h × 2 L h 01 D j ( n ) W 2 L h × L h 10 h ^ ‾ i ( n ) - - - ( 6 )
In formula
D i = diag ( F 2 L h × 2 L h ( x i ( n ) 2 L h × 1 ) ) ;
W L h × 2 L h 01 = F L h × L h 0 L h × L h I L h × L h F 2 L h × 2 L h - 1 ;
W 2 L h × L h 10 = F 2 L h × 2 L h I L h × L h 0 L h × L h T F L h × L h - 1 ;
In formula, i ≠ j; I, j=1,2 ..., M, be respectively L h× L hdimension Fourier transform matrix and Fourier transform inverse matrix; Diag represents diagonal matrix; represent L h× L hdimension unit matrix; represent L h× L hdimension null matrix.
Utilize normalization hyperchannel frequency domain least-square methods (Normalized Multichannel Frequency-domain Least Mean Square, NMCFLMS), its cost function is
J f ( n ) = Σ i = 1 M - 1 Σ j = i + 1 M e ‾ ij H ( n ) e ‾ ij ( n ) - - - ( 7 )
In the time that channel additive noise exists, utilize formula (7) to carry out channel frequency domain response and estimate, may disperse, channel estimating was lost efficacy.For making NMCFLMS method in the time that channel additive noise exists, there is good channel estimating performance, can utilize acoustical signal spectrum energy to there is equally distributed characteristic.For sound spectrum is uniformly distributed, utilize the law of the mean pair frequency band energy corresponding with channel impulse response to retrain, constraint condition is
max J P ( n ) = Σ i = 1 M Σ j = 0 L - 1 ln ( | h ^ ‾ ij | 2 ) s . t . Σ i = 1 M Σ j = 0 L h - 1 | h ^ ‾ ij | 2 = 1 ML h - - - ( 8 )
In formula, J p(n) be the penalty under constraint condition, at penalty J p(n) while maximization, constraint condition set up; S.t. represent constraint condition.
In order to make full use of the advantage of NMCFLMS and to make spectrum energy there is equally distributed characteristic, penalty J p(n) by the cost function J of Lagrange multiplier β (n) to NMCFLMS f(n) revise, the cost function that obtains revising NMCFLMS (MNMCFLMS) is
J mod(n)=J f(n)+β(n){-J p(n)} (9)
In formula
β ( n ) = | ▿ J p H ( n ) ▿ J f ( n ) | | | ▿ J p ( n ) | | 2 - - - ( 10 )
In formula, the cost function gradient that the value of Lagrange multiplier β (n) is revised during by stable state equals zero, i.e. ▽ J f(n)=β (n) ▽ J p(n) time, obtain, H represents conjugate transpose;
By the cost function J revising mod(n), obtain channel frequency domain response more new formula be
h ^ ‾ i 10 ( n + 1 ) = h ^ ‾ i 10 ( n ) - ▿ J f 01 ( n ) + β ^ ( n ) ▿ J p 10 ( n ) - - - ( 11 )
Wherein,
▿ J p 01 ( n ) = F 2 L h × 2 L h 0 L h × L h I L h × L h T K ( n ) F 2 L h × 2 L h 0 L h × L h I L h × L h T h ^ ‾ ( n )
β ^ ( n ) = | ▿ J p 10 ( n ) H ▿ J f 01 ( n ) | | | ▿ J p 10 | | 2
▿ J f 01 ( n ) = μ ( P i ( n ) + δ I 2 L h × L h ) - 1 × Σ k = 1 M D k * ( n ) e ‾ ki 01 ( n )
In above formula,
P i ( n ) = λ P i ( n - 1 ) + ( 1 - λ ) × Σ k = 1 , k ≠ i M D k * ( n ) D k ( n )
e ‾ ki 01 ( n ) = F 2 L h × 2 L h 0 L h × L h I L h × L h T F L h × L h - 1 e ‾ ki ( n )
In formula, K (n) is that diagonal entry is diagonal matrix; P i(n) be the spectrum energy of multi-channel output signal, can obtain more stable spectrum energy by forgetting factor λ, parameter δ is normal number, and it can effectively solve the noise scale-up problem that spectrum energy hour causes;
Revise normalization hyperchannel frequency domain least-square methods, impulse response the calculation delay that can estimate adaptively different array elements are poor, avoid channel estimating to occur worsening;
Step B-5, to channel frequency domain response carry out Fourier inversion, obtain channel impulse response estimation and be
h ^ i ( n ) = I L h × L h 0 L h × L h F 2 L h × 2 L h - 1 h ^ ‾ i 10 ( n ) - - - ( 12 )
In formula, for 2L h× 2L hdimension Fourier transform inverse matrix;
Step B-6, according to the channel impulse response of estimating carry out time delay estimation.In adaptive process, a peak value of middle appearance, the corresponding time delay of peak value is exactly the reception direct sound wave signal time delay of i microphone, and sound-source signal arrives the delay inequality between i microphone and j microphone and is
τ ^ ij = ( max l = 1 L h | h ^ il ( n ) | - max l = 1 L h | h ^ jl ( n ) | ) / f s - - - ( 13 )
In formula, f sfor the sample frequency of signal, max represents to get maximal value;
Step C, is multiplied each other and is obtained sound-source signal and arrive the range difference of each microphone by the delay inequality described in step B and the velocity of sound, and according to the position relationship of each microphone, determine the position of described sound source.
Use d ijrepresent the distance between i and j microphone, the velocity of propagation that c is sound source.In the time that the position coordinates of sound source is (x, y, z), sound-source signal is to the delay inequality τ between i microphone and j microphone ijbe a constant, at this moment
d ij = c · τ ^ ij - - - ( 14 )
Formula (14) comprises following concrete equation:
First group: x 2 + y 2 + z 2 = ( x - d ) 2 + y 2 + z 2 = c · τ ^ 21 - - - ( 15 )
x 2 + y 2 + z 2 = x 2 + ( y - d ) 2 + z 2 = c · τ ^ 23 - - - ( 16 )
Second group: ( x - D ) 2 + y 2 + z 2 = ( x - D + d ) 2 + y 2 + z 2 = c · τ ^ 45 - - - ( 17 )
( x - D ) 2 + y 2 + z 2 - = x 2 + ( y - d ) 2 + z 2 = c · τ ^ 46 - - - ( 18 )
In formula, for the delay inequality of sound-source signal to the second microphone and first microphone, for the delay inequality of sound-source signal to the second microphone and the 3rd microphone, for the delay inequality of four microphones of sound-source signal to the and the 5th microphone, for the delay inequality of four microphones of sound-source signal to the and the 6th microphone.
By formula (15) and formula (16),
y=a 1x+b 1 (19)
In formula
a 1 = τ ^ 23 / τ ^ 21 - - - ( 20 )
b 1 = ( c 2 τ ^ 21 τ ^ 23 + d 2 ) ( τ ^ 21 - τ ^ 23 ) / ( 2 d τ ^ 21 ) - - - ( 21 )
By formula (17) and (18),
y=a 2x+b 2 (22)
In formula
a 2 = - τ ^ 46 / τ ^ 45 - - - ( 23 )
b 2 = ( ( c 2 τ ^ 45 τ ^ 46 + d 2 ) ( τ ^ 45 - τ ^ 46 ) + 2 d D / τ ^ 45 ) ( 2 d τ ^ 45 ) - - - ( 24 )
By formula (19) and (22), obtain sound source and in projection coordinate's estimated value of xoy plane be
x ^ = b 2 - b 1 a 1 - a 2 - - - ( 25 )
y ^ = a 1 b 2 - a 2 b 1 a 1 - a 2 - - - ( 26 )
z ^ = 0 - - - ( 27 )
By formula (25)-Shi (26) substitution formula (15)-Shi (18) respectively, obtain four estimated values of sound source Z coordinate, be designated as respectively accurate estimation with the mean value of these four estimated values as sound source Z coordinate,
z ^ = z ^ 1 + z ^ 2 + z ^ 3 + z ^ 4 4 - - - ( 28 )
Through type (25), (26) and the value that (28) obtain are exactly the coordinate figure of auditory localization.
Embodiment mono-:
In order to verify the performance of the inventive method, we carry out Performance Detection for the present invention.In detection, Fig. 4 is microphone placing mould graphoid in room.The acoustic enviroment of common conference chamber, room-sized is 6m × 5m × 3.5m.Microphone array is erected at overhead on the position of about 0.5m, 6 microphones are arranged in double L-shaped microphone array as shown in Figure 4, microphone is respectively 0.5m, 0.5m from the distance on wall and ground, distance between adjacent two microphones is 1m, and the distance between L-type microphone array is 5m.Consider that microphone is little from wall distance, reverberation large, practical application has little significance, therefore, the dash area in Fig. 4 is not considered, and sound source is positioned over to the arbitrary position within two L-type microphones.Sound-source signal is that one section of male voice of recording is in advance read aloud, and its sample frequency is 25kHz.The position laying according to microphone, the length L of channel impulse response hparameter δ=0.1556 × 10 of=370, MNMCFLMS -5, μ=0.5.
Fig. 5 is SNR=20dB, mixing time delay RT 60when=100ms, normalized projection error (Normalized Projection Misalignment, NPM) convergence curve, projection error NPM is
NPM ( n ) = 20 log 10 ( | | h ( n ) - h ( n ) T h ^ ( n ) h ^ ( n ) T h ^ ( n ) h ^ ( n ) | / | | | h ( n ) | | ) - - - ( 29 )
In formula, h (n) is the true impulse response of channel, for channel self-adapting is estimated the impulse response obtaining.
Fig. 5 shows having under channel additive noise environment, and NMCFLMS method has just been dispersed in the time of iteration approximately 450 times.And the inventive method MNMCFLMS in the time of iteration approximately 450 times with regard to stable convergence, effectively avoid the channel estimating that channel additive noise causes to worsen.
Under different reverberation environment in the time of SNR=20dB, adopt the inventive method MNMCFLMS and classical phase tranformation weighting broad sense cross correlation function method (Generalized Cross Correlation-phase Transform, GCC-PHAT) carry out the contrast of time delay estimated performance, result is as shown in table 1 below:
The performance index of the inventive method and GCC-PHAT method contrast under the low reverberation environment of table 1
In table 1, adopt non-abnormity point number percent (Percentage of Non-abnormal Point, PNP) and root-mean-square error (Root Mean Square Error, the RMSE) index as measure algorithm performance.Wherein, PNP and RMSE are defined as respectively
PNP = ( 1 - 1 N Σ i = 1 N T P ( τ i - τ 0 ) ) × 100 % - - - ( 30 )
RMSE = 1 N Σ i = 1 N ( τ i - τ 0 ) 2 - - - ( 31 )
In formula
T P ( x ) = 0 , | x | ≤ 2 / f s 1 , | x | > 2 / f s - - - ( 32 )
In formula, N represents time delay appreciable amt, τ 0the actual value that time delay is estimated.At this, using the time delay estimated value that differs more than 2 sample point with actual value as an abnormity point, non-abnormal number percent is higher, and method performance is better.Table 1 shows, process many experiments, and under low reverberation environment, the inventive method is suitable with GCC-PHAT method performance; In the time that reverberation is greater than 250ms, the non-dissimilarity number percent of GCC-PHAT method significantly declines, root-mean-square error is much larger than sampling precision, cannot effectively carry out time delay estimation; But the performance of the inventive method is uninfluenced, under the environment in reverberation lower than 700ms, still estimation time delay value effectively.
Under room model, real sound source is tested, detect the performance of sound localization method of the present invention.Indoor environment parameter is as follows: SNR=20dB, RT 60=150ms.The result of auditory localization, adopts absolute error as the judgment criteria of positioning performance, and for the coordinate estimated value of sound source position, the present invention is carried out to many experiments checking, result is as shown in table 2 below:
The test result of table 2 the inventive method
From upper table 2, the absolute error of this method orientation distance γ is stable be distributed in 10cm within, the absolute error of horizontal angle θ is controlled within 1 °, height absolute error is stable be distributed in 10cm within.
The disclosed technological means of the present invention program is not limited only to the disclosed technological means of above-mentioned embodiment, also comprises the technical scheme being made up of above technical characterictic combination in any.It should be pointed out that for those skilled in the art, under the premise without departing from the principles of the invention, can also make some improvements and modifications, these improvements and modifications are also considered as protection scope of the present invention.

Claims (5)

1. a three dimensions sound localization method, is characterized in that, comprises the steps:
Steps A, in three-dimensional cartesian coordinate system, the L-type microphone array of setting up two at grade and being oppositely arranged;
Step B, adopts the normalization hyperchannel frequency domain least-square methods NMCFLMS revising to estimate that sound-source signal arrives the delay inequality of each microphone, comprises the steps:
Step B-1, sound-source signal s (n) is through the channel impulse response h of i microphone i(n) rear and channel additive noise v i(n) merge, obtain the reception signal x of i microphone i(n): x i(n)=s t(n) h i(n)+v i(n);
Wherein, T representing matrix matrix transpose operation; N is time series, is integer;
In the time disregarding channel additive noise, the reception signal x of i microphone i(n) with the reception signal x of j microphone j(n) pass between is:
x i T ( n ) h j ( n ) = x j T ( n ) h i ( n ) ;
Step B-2, with length be 2L hthe reception signal x of rectangular window function w (n) to i microphone i(n) carry out windowing process, the n frame that obtains i microphone receives signal and is:
x i ( n ) 2 L h × 1 = [ x i ( n L h - L h ) , x i ( n L h - L h + 1 ) , . . . , x i ( n L h + L h - 1 ) ] T ;
In formula, L hrepresent channel impulse response h i(n) length is integer; N=1 ..., int (N/L h-1); N is the sequence total length that microphone receives signal, is integer; Int (N/L h-1) be to N/L h-1 integer obtaining after rounding downwards;
Step B-3, utilizes Fourier transform that the time-domain signal obtaining in step B-2 is transformed to frequency-region signal:
X i ( n ) 2 L h × 1 = F 2 L h × 2 L h x i ( n ) 2 L h × 1 ;
In formula, for 2L h× 2L hdimension Fourier transform matrix;
Step B-4, introduces penalty to spectrum energy correction, estimates adaptively channel frequency domain response
As channel additive noise v i(n), while existence, error of frequency domain function definition is
e ‾ ij ( n ) = W L h × 2 L h 01 D i ( n ) W 2 L h × L h 10 h ^ ‾ j ( n ) - W L h × 2 L h 01 D j ( n ) W 2 L h × L h 10 h ^ ‾ i ( n )
In formula
D i = diag ( F 2 L h × 2 L h ( x i ( n ) 2 L h × 1 ) ) ;
W L h × 2 L h 01 = F L h × L h 0 L h × L h I L h × L h F 2 L h × 2 L h - 1 ;
W 2 L h × L h 10 = F 2 L h × 2 L h I L h × L h 0 L h × L h T F L h × L h - 1 ;
In formula, i ≠ j; I, j=1,2 ..., M, be respectively L h× L hdimension Fourier transform matrix and Fourier transform inverse matrix; Diag represents diagonal matrix; represent L h× L hdimension unit matrix; represent L h× L hdimension null matrix;
Utilize penalty J p(n) by the cost function J of Lagrange multiplier β (n) to NMCFLMS f(n) revise, the cost function that obtains revising is
J mod(n)=J f(n)+β(n){-J p(n)}
In formula
β ( n ) = | ▿ J p H ( n ) ▿ J f ( n ) | | | ▿ J p ( n ) | | 2
In formula, the cost function gradient that the value of Lagrange multiplier β (n) is revised during by stable state equals zero, i.e. ▽ J f(n)=β (n) ▽ J p(n) time, obtain, H represents conjugate transpose;
By the cost function J revising mod(n), obtain channel frequency domain response more new formula be
h ^ ‾ i 10 ( n + 1 ) = h ^ ‾ i 10 ( n ) - ▿ J f 01 ( n ) + β ^ ( n ) ▿ J p 10 ( n )
Wherein,
▿ J p 01 ( n ) = F 2 L h × 2 L h 0 L h × L h I L h × L h T K ( n ) F 2 L h × 2 L h 0 L h × L h I L h × L h T h ^ ‾ ( n )
β ^ ( n ) = | ▿ J p 10 ( n ) H ▿ J f 01 ( n ) | | | ▿ J p 10 | | 2
▿ J f 01 ( n ) = μ ( P i ( n ) + δ I 2 L h × L h ) - 1 × Σ k = 1 M D k * ( n ) e ‾ ki 01 ( n )
In above formula,
P i ( n ) = λ P i ( n - 1 ) + ( 1 - λ ) × Σ k = 1 , k ≠ i M D k * ( n ) D k ( n )
e ‾ ki 01 ( n ) = F 2 L h × 2 L h 0 L h × L h I L h × L h T F L h × L h - 1 e ‾ ki ( n )
In formula, K (n) is that diagonal entry is diagonal matrix; P i(n) be the spectrum energy of multi-channel output signal, forgetting factor parameter δ is normal number;
Step B-5, to channel frequency domain response carry out Fourier inversion, obtain channel impulse response estimation and be
h ^ i ( n ) = I L h × L h 0 L h × L h F 2 L h × 2 L h - 1 h ^ ‾ i 10 ( n )
In formula, for 2L h× 2L hdimension Fourier transform inverse matrix;
Step B-6, in adaptive process, the channel impulse response of estimation the corresponding time delay of the peak value of middle appearance is the reception direct sound wave signal time delay of i microphone, and sound-source signal arrives the delay inequality between i microphone and j microphone and is
τ ^ ij = ( max l = 1 L h | h ^ il ( n ) | - max l = 1 L h | h ^ jl ( n ) | ) / f s
In formula, f sfor the sample frequency of signal, max represents to get maximal value;
Step C, is multiplied each other and is obtained sound-source signal and arrive the range difference of each microphone by the delay inequality described in step B and the velocity of sound, and according to the position relationship of each microphone, determine the position of described sound source.
2. three dimensions sound localization method according to claim 1, is characterized in that, described J p(n) be the penalty under constraint condition, constraint condition is:
max J P ( n ) = Σ i = 1 M Σ j = 0 L - 1 ln ( | h ^ ‾ ij | 2 ) s . t . Σ i = 1 M Σ j = 0 L h - 1 | h ^ ‾ ij | 2 = 1 ML h
In formula, at penalty J p(n) while maximization, constraint condition set up; S.t. represent constraint condition.
3. three dimensions sound localization method according to claim 1 and 2, is characterized in that, the length L of channel impulse response in described step B-2 hrestrictive condition is:
L h ≤ 2 f s max i , j = 1 M { τ ij , max }
In formula, τ ij, max=d ij/ c is that between i and j microphone, maximum delay is poor, d ijbe the distance between i and j microphone, the velocity of propagation that c is sound source.
4. three dimensions sound localization method according to claim 1, is characterized in that, in described steps A, to lay rule as follows for two L-type arrays: first microphone mic1 is placed in X-axis, and its coordinate is (d, 0,0); Second microphone mic2 and true origin are overlapping, and its coordinate is (0,0,0); The 3rd microphone mic3 is placed in Y-axis, and its coordinate is (0, d, 0), and d is distance between two adjacent microphones, is real number, and first, second and third microphone has formed first L-type microphone array; Meanwhile, taking x=D/2 as axis of symmetry, set up second the L-type microphone array relative with first L-type microphone array, wherein, the coordinate of the 4th microphone mic4 is (D, 0,0), the coordinate of the 5th microphone mic5 is (D-d, 0,0), the coordinate of the 6th microphone mic6 is (D, d, 0); D is two distances between relative L-type microphone array.
5. according to the three dimensions sound localization method described in claim 1 or 4, it is characterized in that, in described step C, determine that the step of sound source position is as follows:
Step C-1, position is the delay inequality that the sound-source signal of (x, y, z) arrives i microphone and j microphone with i microphone and j microphone space from d ijpass be relation has following two groups in detail:
First group: x 2 + y 2 + z 2 = 1 2 c τ ^ 21 ( c 2 τ ^ 21 2 + 2 dx - d 2 ) With
x 2 + y 2 + z 2 = 1 2 c τ ^ 23 ( c 2 τ ^ 23 2 + 2 dy - d 2 ) ;
Second group: ( x - D ) 2 + y 2 + z 2 = 1 2 c τ ^ 45 ( c 2 τ ^ 45 2 d 2 - 2 d ( x - D ) ) With
( x - D ) 2 + y 2 + z 2 = 1 2 c τ ^ 46 ( c 2 τ ^ 46 2 + 2 Dx + D 2 - 2 dy - d 2 ) ;
In formula, for the delay inequality between sound-source signal to the second microphone and first microphone; for the delay inequality of sound-source signal to the second microphone and the 3rd microphone, for the delay inequality of four microphones of sound-source signal to the and the 5th microphone, for the delay inequality of four microphones of sound-source signal to the and the 6th microphone;
Step C-2, according to first group described in step C-1 with second group of relation, obtain the projection coordinate of sound source in xoy plane and be
x ^ = b 2 - b 1 a 1 - a 2
y ^ = a 1 b 2 - a 2 b 1 a 1 - a 2
z ^ = 0
In formula, a 1 = τ ^ 23 / τ ^ 21 ; a 2 = - τ ^ 46 / τ ^ 45 ; b 1 = ( c 2 τ ^ 21 τ ^ 23 + d 2 ) ( τ ^ 21 - τ ^ 23 ) / ( 2 d τ ^ 21 ) ; b 2 = ( ( c 2 τ ^ 45 τ ^ 46 + d 2 ) ( τ ^ 45 - τ ^ 46 ) + 2 dD τ ^ 45 ) / ( 2 d τ ^ 45 ) ;
Step C-3, projection coordinate by the sound source described in step C-2 in xoy plane in first group described in substitution step C-1 and four relational expressions of second group, obtains four z axial coordinate estimated values of sound source respectively the mean value of getting these four z coordinates obtains the z coordinate estimation of sound source for:
z ^ = z ^ 1 + z ^ 2 + z ^ 3 + z ^ 4 4 .
CN201410202062.0A 2014-05-13 2014-05-13 A kind of sound source localization method of three-dimensional space Expired - Fee Related CN103995252B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201410202062.0A CN103995252B (en) 2014-05-13 2014-05-13 A kind of sound source localization method of three-dimensional space

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201410202062.0A CN103995252B (en) 2014-05-13 2014-05-13 A kind of sound source localization method of three-dimensional space

Publications (2)

Publication Number Publication Date
CN103995252A true CN103995252A (en) 2014-08-20
CN103995252B CN103995252B (en) 2016-08-24

Family

ID=51309468

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201410202062.0A Expired - Fee Related CN103995252B (en) 2014-05-13 2014-05-13 A kind of sound source localization method of three-dimensional space

Country Status (1)

Country Link
CN (1) CN103995252B (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104076331A (en) * 2014-06-18 2014-10-01 南京信息工程大学 Sound source positioning method for seven-element microphone array
CN104535965A (en) * 2014-12-29 2015-04-22 江苏科技大学 Parallelized sound source positioning system based on embedded GPU system and method
CN104977564A (en) * 2015-07-09 2015-10-14 百度在线网络技术(北京)有限公司 Microphone array for household intelligent robot based artificial intelligence
CN105681972A (en) * 2016-01-14 2016-06-15 南京信息工程大学 Linearly constrained minimum variance diagonal loaded robust frequency-invariant beam forming method
CN106842111A (en) * 2016-12-28 2017-06-13 西北工业大学 Indoor sound localization method based on microphone mirror image
CN106886010A (en) * 2017-01-17 2017-06-23 南京航空航天大学 A kind of sound bearing recognition methods based on mini microphone array
CN106950542A (en) * 2016-01-06 2017-07-14 中兴通讯股份有限公司 The localization method of sound source, apparatus and system
CN107144820A (en) * 2017-06-21 2017-09-08 歌尔股份有限公司 Sound localization method and device
CN107340498A (en) * 2016-05-03 2017-11-10 深圳光启合众科技有限公司 The determination method and apparatus of robot and sound source position
CN110068797A (en) * 2019-04-23 2019-07-30 浙江大华技术股份有限公司 A kind of method, sound localization method and relevant device for calibrating microphone array
CN110161454A (en) * 2019-06-14 2019-08-23 哈尔滨工业大学 Signal frequency and two dimension DOA combined estimation method based on double L-shaped array
US10491995B1 (en) 2018-10-11 2019-11-26 Cisco Technology, Inc. Directional audio pickup in collaboration endpoints
CN111157951A (en) * 2020-01-13 2020-05-15 东北大学秦皇岛分校 Three-dimensional sound source positioning method based on differential microphone array
CN111856400A (en) * 2020-07-29 2020-10-30 中北大学 Underwater target sound source positioning method and system
US11076251B2 (en) 2019-11-01 2021-07-27 Cisco Technology, Inc. Audio signal processing based on microphone arrangement
CN116299147A (en) * 2023-03-13 2023-06-23 中国科学院声学研究所 One-dimensional structure internal sound source positioning method based on acoustic coherence technology

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020052685A1 (en) * 2000-10-27 2002-05-02 Tsuyoshi Kamiya Position guiding method and system using sound changes
JP2011059050A (en) * 2009-09-14 2011-03-24 Univ Of Tokyo Apparatus and method for detecting sound source direction
CN102033223A (en) * 2010-12-29 2011-04-27 北京信息科技大学 Method for positioning sound source by using microphone array
CN102707262A (en) * 2012-06-20 2012-10-03 太仓博天网络科技有限公司 Sound localization system based on microphone array
CN102866385A (en) * 2012-09-10 2013-01-09 上海大学 Multi-sound-source locating method based on spherical microphone array
CN103076593A (en) * 2012-12-28 2013-05-01 中国科学院声学研究所 Sound source localization method and device

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20020052685A1 (en) * 2000-10-27 2002-05-02 Tsuyoshi Kamiya Position guiding method and system using sound changes
JP2011059050A (en) * 2009-09-14 2011-03-24 Univ Of Tokyo Apparatus and method for detecting sound source direction
CN102033223A (en) * 2010-12-29 2011-04-27 北京信息科技大学 Method for positioning sound source by using microphone array
CN102707262A (en) * 2012-06-20 2012-10-03 太仓博天网络科技有限公司 Sound localization system based on microphone array
CN102866385A (en) * 2012-09-10 2013-01-09 上海大学 Multi-sound-source locating method based on spherical microphone array
CN103076593A (en) * 2012-12-28 2013-05-01 中国科学院声学研究所 Sound source localization method and device

Cited By (24)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104076331A (en) * 2014-06-18 2014-10-01 南京信息工程大学 Sound source positioning method for seven-element microphone array
CN104535965A (en) * 2014-12-29 2015-04-22 江苏科技大学 Parallelized sound source positioning system based on embedded GPU system and method
CN104977564A (en) * 2015-07-09 2015-10-14 百度在线网络技术(北京)有限公司 Microphone array for household intelligent robot based artificial intelligence
CN106950542A (en) * 2016-01-06 2017-07-14 中兴通讯股份有限公司 The localization method of sound source, apparatus and system
CN105681972B (en) * 2016-01-14 2018-05-01 南京信息工程大学 The constant Beamforming Method of sane frequency that linear constraint minimal variance diagonally loads
CN105681972A (en) * 2016-01-14 2016-06-15 南京信息工程大学 Linearly constrained minimum variance diagonal loaded robust frequency-invariant beam forming method
CN107340498A (en) * 2016-05-03 2017-11-10 深圳光启合众科技有限公司 The determination method and apparatus of robot and sound source position
CN106842111A (en) * 2016-12-28 2017-06-13 西北工业大学 Indoor sound localization method based on microphone mirror image
CN106842111B (en) * 2016-12-28 2019-03-29 西北工业大学 Indoor sound localization method based on microphone mirror image
CN106886010B (en) * 2017-01-17 2019-07-30 南京航空航天大学 A kind of sound bearing recognition methods based on mini microphone array
CN106886010A (en) * 2017-01-17 2017-06-23 南京航空航天大学 A kind of sound bearing recognition methods based on mini microphone array
CN107144820A (en) * 2017-06-21 2017-09-08 歌尔股份有限公司 Sound localization method and device
US10687139B2 (en) 2018-10-11 2020-06-16 Cisco Technology, Inc. Directional audio pickup in collaboration endpoints
US10491995B1 (en) 2018-10-11 2019-11-26 Cisco Technology, Inc. Directional audio pickup in collaboration endpoints
CN110068797A (en) * 2019-04-23 2019-07-30 浙江大华技术股份有限公司 A kind of method, sound localization method and relevant device for calibrating microphone array
CN110068797B (en) * 2019-04-23 2021-02-02 浙江大华技术股份有限公司 Method for calibrating microphone array, sound source positioning method and related equipment
CN110161454A (en) * 2019-06-14 2019-08-23 哈尔滨工业大学 Signal frequency and two dimension DOA combined estimation method based on double L-shaped array
US11076251B2 (en) 2019-11-01 2021-07-27 Cisco Technology, Inc. Audio signal processing based on microphone arrangement
US11399248B2 (en) 2019-11-01 2022-07-26 Cisco Technology, Inc. Audio signal processing based on microphone arrangement
CN111157951A (en) * 2020-01-13 2020-05-15 东北大学秦皇岛分校 Three-dimensional sound source positioning method based on differential microphone array
CN111856400A (en) * 2020-07-29 2020-10-30 中北大学 Underwater target sound source positioning method and system
CN111856400B (en) * 2020-07-29 2021-04-09 中北大学 Underwater target sound source positioning method and system
CN116299147A (en) * 2023-03-13 2023-06-23 中国科学院声学研究所 One-dimensional structure internal sound source positioning method based on acoustic coherence technology
CN116299147B (en) * 2023-03-13 2023-11-28 中国科学院声学研究所 One-dimensional structure internal sound source positioning method based on acoustic coherence technology

Also Published As

Publication number Publication date
CN103995252B (en) 2016-08-24

Similar Documents

Publication Publication Date Title
CN103995252A (en) Three-dimensional space sound source positioning method
CN103308889B (en) Passive sound source two-dimensional DOA (direction of arrival) estimation method under complex environment
CN103439688B (en) Sound source positioning system and method used for distributed microphone arrays
Pavlidi et al. Real-time multiple sound source localization using a circular microphone array based on single-source confidence measures
CN104142492A (en) SRP-PHAT multi-source spatial positioning method
CN107167770B (en) A kind of microphone array sound source locating device under the conditions of reverberation
CN102324237A (en) Microphone array voice wave beam formation method, speech signal processing device and system
Niwa et al. Post-filter design for speech enhancement in various noisy environments
Tellakula Acoustic source localization using time delay estimation
CN105388459A (en) Robustness sound source space positioning method of distributed microphone array network
CN103856877A (en) Sound control information detection method and electronic device
CN109188362A (en) A kind of microphone array auditory localization signal processing method
Pertilä et al. Passive self-localization of microphones using ambient sounds
CN106093920B (en) It is a kind of based on the adaptive beam-forming algorithm diagonally loaded
Gaubitch et al. Calibration of distributed sound acquisition systems using TOA measurements from a moving acoustic source
Saqib et al. Sound-based distance estimation for indoor navigation in the presence of ego noise
Svaizer et al. Environment aware estimation of the orientation of acoustic sources using a line array
Niwa et al. PSD estimation in beamspace using property of M-matrix
CN109254265A (en) A kind of whistle vehicle positioning method based on microphone array
KR20090128221A (en) Method for sound source localization and system thereof
CN101982793B (en) Mobile sound source positioning method based on stereophonic signals
Xu et al. Sound source localization based on improved adaptive beamforming
Dmochowski et al. Fast steered response power source localization using inverse mapping of relative delays
CN109001680A (en) The sparse optimization algorithm of block in auditory localization
Chen et al. A sound source localization device based on rectangular pyramid structure for mobile robot

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20160824

Termination date: 20190513

CF01 Termination of patent right due to non-payment of annual fee