CN103995252A - Three-dimensional space sound source positioning method - Google Patents
Three-dimensional space sound source positioning method Download PDFInfo
- Publication number
- CN103995252A CN103995252A CN201410202062.0A CN201410202062A CN103995252A CN 103995252 A CN103995252 A CN 103995252A CN 201410202062 A CN201410202062 A CN 201410202062A CN 103995252 A CN103995252 A CN 103995252A
- Authority
- CN
- China
- Prior art keywords
- microphone
- tau
- sound
- formula
- overbar
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S5/00—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
- G01S5/18—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using ultrasonic, sonic, or infrasonic waves
- G01S5/22—Position of source determined by co-ordinating a plurality of position lines defined by path-difference measurements
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- Circuit For Audible Band Transducer (AREA)
Abstract
The invention provides a three-dimensional space sound source positioning method, and belongs to the technical field of sound source positioning. The method includes the steps that a double-L-shaped microphone array is established; on the basis of the normalization frequency domain minimum mean square method, a penalty function is introduced to correct frequency spectrum energy, and then impulse responses of different array elements are adaptively estimated and delay inequality is calculated; position coordinates of a sound source are determined through the relation between the acquired delay inequality and the position of double-L-shaped microphones. Compared with a traditional positioning method, the array structure is simple, the calculated quantity is small, anti-noise performance, reverberation resistance and positioning accuracy are effectively improved, and the method is more suitable for indoor three-dimensional sound source positioning and can be widely applied to a car hands-free phone, a video conference system, a speech recognition system, an intelligent robot and other fields.
Description
Technical field
The invention belongs to auditory localization technical field, especially relate to a kind of method of utilizing microphone array to carry out sound source three-dimensional location.
Background technology
At present, auditory localization based on microphone array is a major issue in acoustics signal processing field, compared to traditional Array Signal Processing, the voice signal of array microphone processing does not have carrier wave, the range of signal that can process is wide, adaptable, be all widely used in fields such as vehicle-carried hands-free telephone, video conferencing system, speech recognition system and intelligent robots.Auditory localization based on microphone array mainly contain based on steerable beam form localization method, based on arrive delay inequality localization method and based on high-resolution localization method.Wherein, the voice signal that the localization method forming based on steerable beam receives microphone array carries out filtering, weighted sum, then directly control microphone and point to the direction that makes wave beam have peak power output, because its computation complexity height can not be used to real time processing system; The correlation matrix solving between Mike's signal based on high-resolution localization method utilization is made deflection, thereby further make sound source position, although be successfully applied to the processing of some array signals, but locating effect is not good in actual applications, be not only subject to array structure restriction and also signal not steadily time calculated amount can be multiplied; And wherein, be by estimating that sound source is to the delay inequality between microphone and according to estimating sound source position in the position of microphone based on arriving the sound localization method of delay inequality, be not subject to array structure restriction, calculated amount little.
Sound localization method based on arriving delay inequality mainly divides two parts to complete: time delay is estimated and location.Delay time estimation method mainly contains broad sense cross-correlation method (Generalized Cross Correlation, GCC), minimum mean square error method (Least Mean Square, and self-adaptive features value decomposition method (Adaptive Engenvalue Decomposition, AED) LMS).Wherein, GCC method performance under reverberation environment can decline a lot; LMS method performance is basic suitable with it; AED method is by estimating that double-channel impulse response suppresses reverberation, but it requires double-channel relatively prime, does not contain common zero point.But in actual indoor environment, impulse response length is generally very long, the relatively prime possibility of double-channel is very little, and AED method is no longer applicable.In order to improve relatively prime possibility, AED binary channels method of estimation is generalized to self-adapting multi-channel method of estimation (Adaptive Multichannel by the people such as Y (Arden) Huang, AMC), proposition utilizes normalization hyperchannel frequency domain least-square methods (Normalized Multichannel Frequency-domain Least Mean Square, NMCFLMS) to estimate the impulse response of each array element.The basic thought of NMCFLMS method is, observation signal is divided into continuous block signal, adopt frequency domain normalization minimum mean-square method to carry out channel estimation in frequency domain, though there is the feature of the low and Newton method fast convergence rate of hyperchannel frequency domain least-square methods complexity, but in the time that channel additive noise exists, may disperse, can not effectively carry out channel estimating.
Localization method mainly contains least square method and geometry location method, the former calculation of complex, and to initial value sensitivity; The latter utilizes microphone position parameter and time delay estimated value to determine bi-curved funtcional relationship, and multiple bi-curved intersection points are definite sound source position, calculates simply, but easily occurs not exclusive closed solution, is difficult to effective location.
Summary of the invention
May disperse in the time that channel additive noise exists for sound localization method in prior art, can not effectively carry out channel estimating and be difficult to the defect of effective location sound source, the invention provides a kind of improved three dimensions sound localization method, utilize the method for revising normalization hyperchannel frequency domain least-square methods (MNMCFLMS) and limiting channel impulse response length to carry out channel estimating, effectively improved the precision of auditory localization.
In order to achieve the above object, the invention provides following technical scheme:
A kind of three dimensions sound localization method, comprises the steps:
Steps A, in three-dimensional cartesian coordinate system, the L-type microphone array of setting up two at grade and being oppositely arranged;
Step B, adopts the normalization hyperchannel frequency domain least-square methods of revising to estimate that sound-source signal arrives the delay inequality of each microphone, comprises the steps:
Step B-1, sound-source signal s (n) is through the channel impulse response h of i microphone
i(n) rear and channel additive noise v
i(n) merge, obtain the reception signal x of i microphone
i(n): x
i(n)=s
t(n) h
i(n)+v
i(n);
Wherein, T representing matrix matrix transpose operation; N is time series, is integer;
In the time disregarding channel additive noise, the reception signal x of i microphone
i(n) with the reception signal x of j microphone
j(n) pass between is:
Step B-2, with length be 2L
hthe reception signal x of rectangular window function w (n) to i microphone
i(n) carry out windowing process, the n frame that obtains i microphone receives signal and is:
In formula, L
hrepresent channel impulse response h
i(n) length is integer; N=1 ..., int (N/L
h-1); N is the sequence total length that microphone receives signal, is integer; Int (N/L
h-1) be to N/L
h-1 integer obtaining after rounding downwards;
Step B-3, utilizes Fourier transform that the time-domain signal obtaining in step B-2 is transformed to frequency-region signal:
In formula,
for 2L
h× 2L
hdimension Fourier transform matrix;
Step B-4, introduces penalty to spectrum energy correction, estimates adaptively channel frequency domain response
As channel additive noise v
i(n), while existence, error of frequency domain function definition is
In formula
In formula, i ≠ j; I, j=1,2 ..., M,
be respectively L
h× L
hdimension Fourier transform matrix and Fourier transform inverse matrix; Diag represents diagonal matrix;
represent
dimension unit matrix;
represent L
h× L
hdimension null matrix;
Utilize penalty J
p(n) by the cost function J of Lagrange multiplier β (n) to NMCFLMS
f(n) revise, the cost function that obtains revising is
J
mod(n)=J
f(n)+β(n){-J
p(n)}
In formula
In formula, the cost function gradient that the value of Lagrange multiplier β (n) is revised during by stable state equals zero, i.e. ▽ J
f(n)=β (n) ▽ J
p(n) time, obtain, H represents conjugate transpose;
By the cost function J revising
mod(n), obtain channel frequency domain response
more new formula be
Wherein,
In above formula,
In formula, K (n) is that diagonal entry is
diagonal matrix; P
i(n) be the spectrum energy of multi-channel output signal, can obtain more stable spectrum energy by forgetting factor λ,
parameter δ is normal number, and it can effectively solve the noise scale-up problem that spectrum energy hour causes;
Step B-5, to channel frequency domain response
carry out Fourier inversion, obtain channel impulse response estimation and be
In formula,
for 2L
h× 2L
hdimension Fourier transform inverse matrix;
Step B-6, in adaptive process, the channel impulse response of estimation
the corresponding time delay of the peak value of middle appearance is exactly the reception direct sound wave signal time delay of i microphone, and sound-source signal arrives the delay inequality between i microphone and j microphone and is
In formula, f
sfor the sample frequency of signal, max represents to get maximal value;
Step C, is multiplied each other and is obtained sound-source signal and arrive the range difference of each microphone by the delay inequality described in step B and the velocity of sound, and according to the position relationship of each microphone, determine the position of described sound source.
Further, described J
p(n) be the penalty under constraint condition, constraint condition is
In formula, at penalty J
p(n) while maximization, constraint condition
set up; S.t. represent constraint condition.
Further, the length L of channel impulse response in described step B-2
hrestrictive condition is:
In formula, τ
ij, max=d
ij/ c is that between i and j microphone, maximum delay is poor, d
ijbe the distance between i and j microphone, the velocity of propagation that c is sound source.
Further, in described steps A two L-type arrays to lay rule as follows: first microphone mic1 is placed in X-axis, and its coordinate is (d, 0,0); Second microphone mic2 and true origin are overlapping, and its coordinate is (0,0,0); The 3rd microphone mic3 is placed in Y-axis, and its coordinate is (0, d, 0), and d is distance between two adjacent microphones, is real number, and first, second and third microphone has formed first L-type microphone array; Meanwhile, taking x=D/2 as axis of symmetry, set up second the L-type microphone array relative with first L-type microphone array, wherein, the coordinate of the 4th microphone mic4 is (D, 0,0), the coordinate of the 5th microphone mic5 is (D-d, 0,0), the coordinate of the 6th microphone mic6 is (D, d, 0); D is two distances between relative L-type microphone array.
Further, in described step C, determine that the step of sound source position is as follows:
Step C-1, position is the delay inequality that the sound-source signal of (x, y, z) arrives i microphone and j microphone
with i microphone and j microphone space from d
ijpass be
relation has following two groups in detail:
First group:
With
Second group:
With
In formula,
for the delay inequality between sound-source signal to the second microphone and first microphone;
for the delay inequality of sound-source signal to the second microphone and the 3rd microphone,
for the delay inequality of four microphones of sound-source signal to the and the 5th microphone,
for the delay inequality of four microphones of sound-source signal to the and the 6th microphone;
Step C-2, according to first group described in step C-1 with second group of relation, obtain the projection coordinate of sound source in xoy plane and be
In formula,
Step C-3, projection coordinate by the sound source described in step C-2 in xoy plane in first group described in substitution step C-1 and four relational expressions of second group, obtains four z axial coordinate estimated values of sound source respectively
getting the mean value of these four z coordinates estimates as the z coordinate of sound source
for:
Compared with prior art, tool of the present invention has the following advantages and beneficial effect:
The present invention, on normalization frequency domain least-square methods basis, by introducing penalty, spectrum energy is revised, thereby it is poor to estimate adaptively impulse response the calculation delay of different array elements, avoids channel estimating to occur worsening; Utilize interior two the unique definite sound sources of crossing straight line of plane projected position planar, avoided the appearance of not exclusive closed solution problem; Define in addition channel impulse response length, strengthened the relatively prime property of double-channel, improved anti-reverberation ability.Compared with traditional localization method, array structure of the present invention is simple, calculated amount is few, noise robustness, anti-reverberation ability and positioning precision are effectively improved, more be suitable for indoor three-dimensional auditory localization, can be widely used in the every field such as vehicle-carried hands-free telephone, video conferencing system, speech recognition system and intelligent robot.
Brief description of the drawings
Fig. 1 is the flow chart of steps of three dimensions sound localization method provided by the invention;
Fig. 2 is double L-shaped microphone array setting principle figure;
Fig. 3 is the normalization hyperchannel frequency domain least-square methods schematic diagram of revising;
Fig. 4 is that in embodiment, microphone array is listed in the schematic diagram of putting in room;
Fig. 5 is normalized projection error convergence curve figure.
Embodiment
Below with reference to specific embodiment, technical scheme provided by the invention is elaborated, should understands following embodiment and only be not used in and limit the scope of the invention for the present invention is described.
First the present invention improves normalization hyperchannel frequency domain least-square methods NMCFLMS, increases its performance; Limit channel impulse response length simultaneously, strengthen the relatively prime property of double-channel; And utilize the NMCFLMS revising to carry out channel estimating.Specifically, process flow diagram as shown in Figure 1, comprises the steps:
Steps A, sets up three-dimensional coordinate system, at grade, 6 microphones is put into two relative L-type arrays, and between them, distance is D, and as shown in Figure 2: first microphone mic1 is placed in X-axis, its coordinate is (d, 0,0); Second microphone mic2 and true origin are overlapping, and its coordinate is (0,0,0); The 3rd microphone mic3 is placed in Y-axis, and its coordinate is (0, d, 0), and d is distance between two adjacent microphones, is real number, and first, second and third microphone has formed first L-type microphone array.Meanwhile, taking x=D/2 as axis of symmetry, set up second the L-type microphone array relative with first L-type microphone array, wherein, the coordinate of the 4th microphone mic4 is (D, 0,0), the coordinate of the 5th microphone mic5 is (D-d, 0,0), the coordinate of the 6th microphone mic6 is (D, d, 0); D is two distances between relative L-type microphone array.The projection coordinate of the coordinate (x, y, z) of sound source in xoy plane is (x, y, 0), and the distance between projection coordinate's point and initial point is γ, and both lines and x axle forward angle are θ.
Step B, adopts the normalization hyperchannel frequency domain least-square methods MNMCFLMS revising, and estimates that sound-source signal arrives the delay inequality of each microphone, its schematic diagram as shown in Figure 3:
Step B-1, the channel impulse response h of sound-source signal s (n), an i microphone
i(n), channel additive noise v
iand the reception signal x of i microphone (n)
i(n) pass between is
x
i(n)=s
T(n)h
i(n)+v
i(n) (1)
In formula, T representing matrix matrix transpose operation; N is time series, is integer.
In the time disregarding channel additive noise, the reception signal x of i microphone
i(n) with the reception signal x of j microphone
j(n) pass between is
In formula, i, j=1,2 ..., M, the number that M is microphone, is positive integer;
Step B-2, carries out frame to the reception signal of i microphone and moves as L
hwindowing process: with length be 2L
hrectangular window function w (n) is multiplied by the reception signal of i microphone, and the n frame that obtains i microphone receives signal and is:
In formula, L
hrepresent channel impulse response h
ilength, be integer; N=1 ..., int (N/L
h-1); N is the sequence total length that microphone receives signal, is integer; Int (N/L
h-1) be to N/L
h-1 integer obtaining after rounding downwards;
In actual indoor environment, in order to overcome the very long very little deficiency of the relatively prime possibility of double-channel that causes of channel impulse response length, guaranteeing that double-channel is relatively prime when stronger, can be further to channel impulse response length L
hlimit, restrictive condition is:
In formula,
be that between i and j microphone, maximum delay is poor.By limiting channel impulse response length, effectively strengthen the relatively prime property of double-channel.
Step B-3, obtains frequency-region signal to formula (3) as Fourier transform
In formula,
for 2L
h× 2L
hdimension Fourier transform matrix.
Step B-4, introduces penalty to spectrum energy correction, estimates adaptively and estimates adaptively channel frequency domain response
As channel additive noise v
i(n), while existence, error of frequency domain function definition is
In formula
In formula, i ≠ j; I, j=1,2 ..., M,
be respectively L
h× L
hdimension Fourier transform matrix and Fourier transform inverse matrix; Diag represents diagonal matrix;
represent L
h× L
hdimension unit matrix;
represent L
h× L
hdimension null matrix.
Utilize normalization hyperchannel frequency domain least-square methods (Normalized Multichannel Frequency-domain Least Mean Square, NMCFLMS), its cost function is
In the time that channel additive noise exists, utilize formula (7) to carry out channel frequency domain response and estimate, may disperse, channel estimating was lost efficacy.For making NMCFLMS method in the time that channel additive noise exists, there is good channel estimating performance, can utilize acoustical signal spectrum energy to there is equally distributed characteristic.For sound spectrum is uniformly distributed, utilize the law of the mean pair frequency band energy corresponding with channel impulse response to retrain, constraint condition is
In formula, J
p(n) be the penalty under constraint condition, at penalty J
p(n) while maximization, constraint condition
set up; S.t. represent constraint condition.
In order to make full use of the advantage of NMCFLMS and to make spectrum energy there is equally distributed characteristic, penalty J
p(n) by the cost function J of Lagrange multiplier β (n) to NMCFLMS
f(n) revise, the cost function that obtains revising NMCFLMS (MNMCFLMS) is
J
mod(n)=J
f(n)+β(n){-J
p(n)} (9)
In formula
In formula, the cost function gradient that the value of Lagrange multiplier β (n) is revised during by stable state equals zero, i.e. ▽ J
f(n)=β (n) ▽ J
p(n) time, obtain, H represents conjugate transpose;
By the cost function J revising
mod(n), obtain channel frequency domain response
more new formula be
Wherein,
In above formula,
In formula, K (n) is that diagonal entry is
diagonal matrix; P
i(n) be the spectrum energy of multi-channel output signal, can obtain more stable spectrum energy by forgetting factor λ,
parameter δ is normal number, and it can effectively solve the noise scale-up problem that spectrum energy hour causes;
Revise normalization hyperchannel frequency domain least-square methods, impulse response the calculation delay that can estimate adaptively different array elements are poor, avoid channel estimating to occur worsening;
Step B-5, to channel frequency domain response
carry out Fourier inversion, obtain channel impulse response estimation and be
In formula,
for 2L
h× 2L
hdimension Fourier transform inverse matrix;
Step B-6, according to the channel impulse response of estimating
carry out time delay estimation.In adaptive process,
a peak value of middle appearance, the corresponding time delay of peak value is exactly the reception direct sound wave signal time delay of i microphone, and sound-source signal arrives the delay inequality between i microphone and j microphone and is
In formula, f
sfor the sample frequency of signal, max represents to get maximal value;
Step C, is multiplied each other and is obtained sound-source signal and arrive the range difference of each microphone by the delay inequality described in step B and the velocity of sound, and according to the position relationship of each microphone, determine the position of described sound source.
Use d
ijrepresent the distance between i and j microphone, the velocity of propagation that c is sound source.In the time that the position coordinates of sound source is (x, y, z), sound-source signal is to the delay inequality τ between i microphone and j microphone
ijbe a constant, at this moment
Formula (14) comprises following concrete equation:
First group:
Second group:
In formula,
for the delay inequality of sound-source signal to the second microphone and first microphone,
for the delay inequality of sound-source signal to the second microphone and the 3rd microphone,
for the delay inequality of four microphones of sound-source signal to the and the 5th microphone,
for the delay inequality of four microphones of sound-source signal to the and the 6th microphone.
By formula (15) and formula (16),
y=a
1x+b
1 (19)
In formula
By formula (17) and (18),
y=a
2x+b
2 (22)
In formula
By formula (19) and (22), obtain sound source and in projection coordinate's estimated value of xoy plane be
By formula (25)-Shi (26) substitution formula (15)-Shi (18) respectively, obtain four estimated values of sound source Z coordinate, be designated as respectively
accurate estimation with the mean value of these four estimated values as sound source Z coordinate,
Through type (25), (26) and the value that (28) obtain are exactly the coordinate figure of auditory localization.
Embodiment mono-:
In order to verify the performance of the inventive method, we carry out Performance Detection for the present invention.In detection, Fig. 4 is microphone placing mould graphoid in room.The acoustic enviroment of common conference chamber, room-sized is 6m × 5m × 3.5m.Microphone array is erected at overhead on the position of about 0.5m, 6 microphones are arranged in double L-shaped microphone array as shown in Figure 4, microphone is respectively 0.5m, 0.5m from the distance on wall and ground, distance between adjacent two microphones is 1m, and the distance between L-type microphone array is 5m.Consider that microphone is little from wall distance, reverberation large, practical application has little significance, therefore, the dash area in Fig. 4 is not considered, and sound source is positioned over to the arbitrary position within two L-type microphones.Sound-source signal is that one section of male voice of recording is in advance read aloud, and its sample frequency is 25kHz.The position laying according to microphone, the length L of channel impulse response
hparameter δ=0.1556 × 10 of=370, MNMCFLMS
-5, μ=0.5.
Fig. 5 is SNR=20dB, mixing time delay RT
60when=100ms, normalized projection error (Normalized Projection Misalignment, NPM) convergence curve, projection error NPM is
In formula, h (n) is the true impulse response of channel,
for channel self-adapting is estimated the impulse response obtaining.
Fig. 5 shows having under channel additive noise environment, and NMCFLMS method has just been dispersed in the time of iteration approximately 450 times.And the inventive method MNMCFLMS in the time of iteration approximately 450 times with regard to stable convergence, effectively avoid the channel estimating that channel additive noise causes to worsen.
Under different reverberation environment in the time of SNR=20dB, adopt the inventive method MNMCFLMS and classical phase tranformation weighting broad sense cross correlation function method (Generalized Cross Correlation-phase Transform, GCC-PHAT) carry out the contrast of time delay estimated performance, result is as shown in table 1 below:
The performance index of the inventive method and GCC-PHAT method contrast under the low reverberation environment of table 1
In table 1, adopt non-abnormity point number percent (Percentage of Non-abnormal Point, PNP) and root-mean-square error (Root Mean Square Error, the RMSE) index as measure algorithm performance.Wherein, PNP and RMSE are defined as respectively
In formula
In formula, N represents time delay appreciable amt, τ
0the actual value that time delay is estimated.At this, using the time delay estimated value that differs more than 2 sample point with actual value as an abnormity point, non-abnormal number percent is higher, and method performance is better.Table 1 shows, process many experiments, and under low reverberation environment, the inventive method is suitable with GCC-PHAT method performance; In the time that reverberation is greater than 250ms, the non-dissimilarity number percent of GCC-PHAT method significantly declines, root-mean-square error is much larger than sampling precision, cannot effectively carry out time delay estimation; But the performance of the inventive method is uninfluenced, under the environment in reverberation lower than 700ms, still estimation time delay value effectively.
Under room model, real sound source is tested, detect the performance of sound localization method of the present invention.Indoor environment parameter is as follows: SNR=20dB, RT
60=150ms.The result of auditory localization, adopts absolute error
as the judgment criteria of positioning performance, and
for the coordinate estimated value of sound source position, the present invention is carried out to many experiments checking, result is as shown in table 2 below:
The test result of table 2 the inventive method
From upper table 2, the absolute error of this method orientation distance γ is stable be distributed in 10cm within, the absolute error of horizontal angle θ is controlled within 1 °, height absolute error is stable be distributed in 10cm within.
The disclosed technological means of the present invention program is not limited only to the disclosed technological means of above-mentioned embodiment, also comprises the technical scheme being made up of above technical characterictic combination in any.It should be pointed out that for those skilled in the art, under the premise without departing from the principles of the invention, can also make some improvements and modifications, these improvements and modifications are also considered as protection scope of the present invention.
Claims (5)
1. a three dimensions sound localization method, is characterized in that, comprises the steps:
Steps A, in three-dimensional cartesian coordinate system, the L-type microphone array of setting up two at grade and being oppositely arranged;
Step B, adopts the normalization hyperchannel frequency domain least-square methods NMCFLMS revising to estimate that sound-source signal arrives the delay inequality of each microphone, comprises the steps:
Step B-1, sound-source signal s (n) is through the channel impulse response h of i microphone
i(n) rear and channel additive noise v
i(n) merge, obtain the reception signal x of i microphone
i(n): x
i(n)=s
t(n) h
i(n)+v
i(n);
Wherein, T representing matrix matrix transpose operation; N is time series, is integer;
In the time disregarding channel additive noise, the reception signal x of i microphone
i(n) with the reception signal x of j microphone
j(n) pass between is:
Step B-2, with length be 2L
hthe reception signal x of rectangular window function w (n) to i microphone
i(n) carry out windowing process, the n frame that obtains i microphone receives signal and is:
In formula, L
hrepresent channel impulse response h
i(n) length is integer; N=1 ..., int (N/L
h-1); N is the sequence total length that microphone receives signal, is integer; Int (N/L
h-1) be to N/L
h-1 integer obtaining after rounding downwards;
Step B-3, utilizes Fourier transform that the time-domain signal obtaining in step B-2 is transformed to frequency-region signal:
In formula,
for 2L
h× 2L
hdimension Fourier transform matrix;
Step B-4, introduces penalty to spectrum energy correction, estimates adaptively channel frequency domain response
As channel additive noise v
i(n), while existence, error of frequency domain function definition is
In formula
In formula, i ≠ j; I, j=1,2 ..., M,
be respectively L
h× L
hdimension Fourier transform matrix and Fourier transform inverse matrix; Diag represents diagonal matrix;
represent L
h× L
hdimension unit matrix;
represent L
h× L
hdimension null matrix;
Utilize penalty J
p(n) by the cost function J of Lagrange multiplier β (n) to NMCFLMS
f(n) revise, the cost function that obtains revising is
J
mod(n)=J
f(n)+β(n){-J
p(n)}
In formula
In formula, the cost function gradient that the value of Lagrange multiplier β (n) is revised during by stable state equals zero, i.e. ▽ J
f(n)=β (n) ▽ J
p(n) time, obtain, H represents conjugate transpose;
By the cost function J revising
mod(n), obtain channel frequency domain response
more new formula be
Wherein,
In above formula,
In formula, K (n) is that diagonal entry is
diagonal matrix; P
i(n) be the spectrum energy of multi-channel output signal, forgetting factor
parameter δ is normal number;
Step B-5, to channel frequency domain response
carry out Fourier inversion, obtain channel impulse response estimation and be
In formula,
for 2L
h× 2L
hdimension Fourier transform inverse matrix;
Step B-6, in adaptive process, the channel impulse response of estimation
the corresponding time delay of the peak value of middle appearance is the reception direct sound wave signal time delay of i microphone, and sound-source signal arrives the delay inequality between i microphone and j microphone and is
In formula, f
sfor the sample frequency of signal, max represents to get maximal value;
Step C, is multiplied each other and is obtained sound-source signal and arrive the range difference of each microphone by the delay inequality described in step B and the velocity of sound, and according to the position relationship of each microphone, determine the position of described sound source.
2. three dimensions sound localization method according to claim 1, is characterized in that, described J
p(n) be the penalty under constraint condition, constraint condition is:
In formula, at penalty J
p(n) while maximization, constraint condition
set up; S.t. represent constraint condition.
3. three dimensions sound localization method according to claim 1 and 2, is characterized in that, the length L of channel impulse response in described step B-2
hrestrictive condition is:
In formula, τ
ij, max=d
ij/ c is that between i and j microphone, maximum delay is poor, d
ijbe the distance between i and j microphone, the velocity of propagation that c is sound source.
4. three dimensions sound localization method according to claim 1, is characterized in that, in described steps A, to lay rule as follows for two L-type arrays: first microphone mic1 is placed in X-axis, and its coordinate is (d, 0,0); Second microphone mic2 and true origin are overlapping, and its coordinate is (0,0,0); The 3rd microphone mic3 is placed in Y-axis, and its coordinate is (0, d, 0), and d is distance between two adjacent microphones, is real number, and first, second and third microphone has formed first L-type microphone array; Meanwhile, taking x=D/2 as axis of symmetry, set up second the L-type microphone array relative with first L-type microphone array, wherein, the coordinate of the 4th microphone mic4 is (D, 0,0), the coordinate of the 5th microphone mic5 is (D-d, 0,0), the coordinate of the 6th microphone mic6 is (D, d, 0); D is two distances between relative L-type microphone array.
5. according to the three dimensions sound localization method described in claim 1 or 4, it is characterized in that, in described step C, determine that the step of sound source position is as follows:
Step C-1, position is the delay inequality that the sound-source signal of (x, y, z) arrives i microphone and j microphone
with i microphone and j microphone space from d
ijpass be
relation has following two groups in detail:
First group:
With
Second group:
With
In formula,
for the delay inequality between sound-source signal to the second microphone and first microphone;
for the delay inequality of sound-source signal to the second microphone and the 3rd microphone,
for the delay inequality of four microphones of sound-source signal to the and the 5th microphone,
for the delay inequality of four microphones of sound-source signal to the and the 6th microphone;
Step C-2, according to first group described in step C-1 with second group of relation, obtain the projection coordinate of sound source in xoy plane and be
In formula,
Step C-3, projection coordinate by the sound source described in step C-2 in xoy plane in first group described in substitution step C-1 and four relational expressions of second group, obtains four z axial coordinate estimated values of sound source respectively
the mean value of getting these four z coordinates obtains the z coordinate estimation of sound source
for:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410202062.0A CN103995252B (en) | 2014-05-13 | 2014-05-13 | A kind of sound source localization method of three-dimensional space |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201410202062.0A CN103995252B (en) | 2014-05-13 | 2014-05-13 | A kind of sound source localization method of three-dimensional space |
Publications (2)
Publication Number | Publication Date |
---|---|
CN103995252A true CN103995252A (en) | 2014-08-20 |
CN103995252B CN103995252B (en) | 2016-08-24 |
Family
ID=51309468
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201410202062.0A Expired - Fee Related CN103995252B (en) | 2014-05-13 | 2014-05-13 | A kind of sound source localization method of three-dimensional space |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103995252B (en) |
Cited By (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104076331A (en) * | 2014-06-18 | 2014-10-01 | 南京信息工程大学 | Sound source positioning method for seven-element microphone array |
CN104535965A (en) * | 2014-12-29 | 2015-04-22 | 江苏科技大学 | Parallelized sound source positioning system based on embedded GPU system and method |
CN104977564A (en) * | 2015-07-09 | 2015-10-14 | 百度在线网络技术(北京)有限公司 | Microphone array for household intelligent robot based artificial intelligence |
CN105681972A (en) * | 2016-01-14 | 2016-06-15 | 南京信息工程大学 | Linearly constrained minimum variance diagonal loaded robust frequency-invariant beam forming method |
CN106842111A (en) * | 2016-12-28 | 2017-06-13 | 西北工业大学 | Indoor sound localization method based on microphone mirror image |
CN106886010A (en) * | 2017-01-17 | 2017-06-23 | 南京航空航天大学 | A kind of sound bearing recognition methods based on mini microphone array |
CN106950542A (en) * | 2016-01-06 | 2017-07-14 | 中兴通讯股份有限公司 | The localization method of sound source, apparatus and system |
CN107144820A (en) * | 2017-06-21 | 2017-09-08 | 歌尔股份有限公司 | Sound localization method and device |
CN107340498A (en) * | 2016-05-03 | 2017-11-10 | 深圳光启合众科技有限公司 | The determination method and apparatus of robot and sound source position |
CN110068797A (en) * | 2019-04-23 | 2019-07-30 | 浙江大华技术股份有限公司 | A kind of method, sound localization method and relevant device for calibrating microphone array |
CN110161454A (en) * | 2019-06-14 | 2019-08-23 | 哈尔滨工业大学 | Signal frequency and two dimension DOA combined estimation method based on double L-shaped array |
US10491995B1 (en) | 2018-10-11 | 2019-11-26 | Cisco Technology, Inc. | Directional audio pickup in collaboration endpoints |
CN111157951A (en) * | 2020-01-13 | 2020-05-15 | 东北大学秦皇岛分校 | Three-dimensional sound source positioning method based on differential microphone array |
CN111856400A (en) * | 2020-07-29 | 2020-10-30 | 中北大学 | Underwater target sound source positioning method and system |
US11076251B2 (en) | 2019-11-01 | 2021-07-27 | Cisco Technology, Inc. | Audio signal processing based on microphone arrangement |
CN116299147A (en) * | 2023-03-13 | 2023-06-23 | 中国科学院声学研究所 | One-dimensional structure internal sound source positioning method based on acoustic coherence technology |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020052685A1 (en) * | 2000-10-27 | 2002-05-02 | Tsuyoshi Kamiya | Position guiding method and system using sound changes |
JP2011059050A (en) * | 2009-09-14 | 2011-03-24 | Univ Of Tokyo | Apparatus and method for detecting sound source direction |
CN102033223A (en) * | 2010-12-29 | 2011-04-27 | 北京信息科技大学 | Method for positioning sound source by using microphone array |
CN102707262A (en) * | 2012-06-20 | 2012-10-03 | 太仓博天网络科技有限公司 | Sound localization system based on microphone array |
CN102866385A (en) * | 2012-09-10 | 2013-01-09 | 上海大学 | Multi-sound-source locating method based on spherical microphone array |
CN103076593A (en) * | 2012-12-28 | 2013-05-01 | 中国科学院声学研究所 | Sound source localization method and device |
-
2014
- 2014-05-13 CN CN201410202062.0A patent/CN103995252B/en not_active Expired - Fee Related
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020052685A1 (en) * | 2000-10-27 | 2002-05-02 | Tsuyoshi Kamiya | Position guiding method and system using sound changes |
JP2011059050A (en) * | 2009-09-14 | 2011-03-24 | Univ Of Tokyo | Apparatus and method for detecting sound source direction |
CN102033223A (en) * | 2010-12-29 | 2011-04-27 | 北京信息科技大学 | Method for positioning sound source by using microphone array |
CN102707262A (en) * | 2012-06-20 | 2012-10-03 | 太仓博天网络科技有限公司 | Sound localization system based on microphone array |
CN102866385A (en) * | 2012-09-10 | 2013-01-09 | 上海大学 | Multi-sound-source locating method based on spherical microphone array |
CN103076593A (en) * | 2012-12-28 | 2013-05-01 | 中国科学院声学研究所 | Sound source localization method and device |
Cited By (24)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104076331A (en) * | 2014-06-18 | 2014-10-01 | 南京信息工程大学 | Sound source positioning method for seven-element microphone array |
CN104535965A (en) * | 2014-12-29 | 2015-04-22 | 江苏科技大学 | Parallelized sound source positioning system based on embedded GPU system and method |
CN104977564A (en) * | 2015-07-09 | 2015-10-14 | 百度在线网络技术(北京)有限公司 | Microphone array for household intelligent robot based artificial intelligence |
CN106950542A (en) * | 2016-01-06 | 2017-07-14 | 中兴通讯股份有限公司 | The localization method of sound source, apparatus and system |
CN105681972B (en) * | 2016-01-14 | 2018-05-01 | 南京信息工程大学 | The constant Beamforming Method of sane frequency that linear constraint minimal variance diagonally loads |
CN105681972A (en) * | 2016-01-14 | 2016-06-15 | 南京信息工程大学 | Linearly constrained minimum variance diagonal loaded robust frequency-invariant beam forming method |
CN107340498A (en) * | 2016-05-03 | 2017-11-10 | 深圳光启合众科技有限公司 | The determination method and apparatus of robot and sound source position |
CN106842111A (en) * | 2016-12-28 | 2017-06-13 | 西北工业大学 | Indoor sound localization method based on microphone mirror image |
CN106842111B (en) * | 2016-12-28 | 2019-03-29 | 西北工业大学 | Indoor sound localization method based on microphone mirror image |
CN106886010B (en) * | 2017-01-17 | 2019-07-30 | 南京航空航天大学 | A kind of sound bearing recognition methods based on mini microphone array |
CN106886010A (en) * | 2017-01-17 | 2017-06-23 | 南京航空航天大学 | A kind of sound bearing recognition methods based on mini microphone array |
CN107144820A (en) * | 2017-06-21 | 2017-09-08 | 歌尔股份有限公司 | Sound localization method and device |
US10687139B2 (en) | 2018-10-11 | 2020-06-16 | Cisco Technology, Inc. | Directional audio pickup in collaboration endpoints |
US10491995B1 (en) | 2018-10-11 | 2019-11-26 | Cisco Technology, Inc. | Directional audio pickup in collaboration endpoints |
CN110068797A (en) * | 2019-04-23 | 2019-07-30 | 浙江大华技术股份有限公司 | A kind of method, sound localization method and relevant device for calibrating microphone array |
CN110068797B (en) * | 2019-04-23 | 2021-02-02 | 浙江大华技术股份有限公司 | Method for calibrating microphone array, sound source positioning method and related equipment |
CN110161454A (en) * | 2019-06-14 | 2019-08-23 | 哈尔滨工业大学 | Signal frequency and two dimension DOA combined estimation method based on double L-shaped array |
US11076251B2 (en) | 2019-11-01 | 2021-07-27 | Cisco Technology, Inc. | Audio signal processing based on microphone arrangement |
US11399248B2 (en) | 2019-11-01 | 2022-07-26 | Cisco Technology, Inc. | Audio signal processing based on microphone arrangement |
CN111157951A (en) * | 2020-01-13 | 2020-05-15 | 东北大学秦皇岛分校 | Three-dimensional sound source positioning method based on differential microphone array |
CN111856400A (en) * | 2020-07-29 | 2020-10-30 | 中北大学 | Underwater target sound source positioning method and system |
CN111856400B (en) * | 2020-07-29 | 2021-04-09 | 中北大学 | Underwater target sound source positioning method and system |
CN116299147A (en) * | 2023-03-13 | 2023-06-23 | 中国科学院声学研究所 | One-dimensional structure internal sound source positioning method based on acoustic coherence technology |
CN116299147B (en) * | 2023-03-13 | 2023-11-28 | 中国科学院声学研究所 | One-dimensional structure internal sound source positioning method based on acoustic coherence technology |
Also Published As
Publication number | Publication date |
---|---|
CN103995252B (en) | 2016-08-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103995252A (en) | Three-dimensional space sound source positioning method | |
CN103308889B (en) | Passive sound source two-dimensional DOA (direction of arrival) estimation method under complex environment | |
CN103439688B (en) | Sound source positioning system and method used for distributed microphone arrays | |
Pavlidi et al. | Real-time multiple sound source localization using a circular microphone array based on single-source confidence measures | |
CN104142492A (en) | SRP-PHAT multi-source spatial positioning method | |
CN107167770B (en) | A kind of microphone array sound source locating device under the conditions of reverberation | |
CN102324237A (en) | Microphone array voice wave beam formation method, speech signal processing device and system | |
Niwa et al. | Post-filter design for speech enhancement in various noisy environments | |
Tellakula | Acoustic source localization using time delay estimation | |
CN105388459A (en) | Robustness sound source space positioning method of distributed microphone array network | |
CN103856877A (en) | Sound control information detection method and electronic device | |
CN109188362A (en) | A kind of microphone array auditory localization signal processing method | |
Pertilä et al. | Passive self-localization of microphones using ambient sounds | |
CN106093920B (en) | It is a kind of based on the adaptive beam-forming algorithm diagonally loaded | |
Gaubitch et al. | Calibration of distributed sound acquisition systems using TOA measurements from a moving acoustic source | |
Saqib et al. | Sound-based distance estimation for indoor navigation in the presence of ego noise | |
Svaizer et al. | Environment aware estimation of the orientation of acoustic sources using a line array | |
Niwa et al. | PSD estimation in beamspace using property of M-matrix | |
CN109254265A (en) | A kind of whistle vehicle positioning method based on microphone array | |
KR20090128221A (en) | Method for sound source localization and system thereof | |
CN101982793B (en) | Mobile sound source positioning method based on stereophonic signals | |
Xu et al. | Sound source localization based on improved adaptive beamforming | |
Dmochowski et al. | Fast steered response power source localization using inverse mapping of relative delays | |
CN109001680A (en) | The sparse optimization algorithm of block in auditory localization | |
Chen et al. | A sound source localization device based on rectangular pyramid structure for mobile robot |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
C14 | Grant of patent or utility model | ||
GR01 | Patent grant | ||
CF01 | Termination of patent right due to non-payment of annual fee |
Granted publication date: 20160824 Termination date: 20190513 |
|
CF01 | Termination of patent right due to non-payment of annual fee |