JP2003078988A - Sound pickup device, method and program, recording medium - Google Patents

Sound pickup device, method and program, recording medium

Info

Publication number
JP2003078988A
JP2003078988A JP2001269751A JP2001269751A JP2003078988A JP 2003078988 A JP2003078988 A JP 2003078988A JP 2001269751 A JP2001269751 A JP 2001269751A JP 2001269751 A JP2001269751 A JP 2001269751A JP 2003078988 A JP2003078988 A JP 2003078988A
Authority
JP
Japan
Prior art keywords
frequency band
signal
sound
level
phase
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
JP2001269751A
Other languages
Japanese (ja)
Other versions
JP3716918B2 (en
Inventor
Mariko Aoki
真理子 青木
Kenichi Furuya
賢一 古家
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nippon Telegraph and Telephone Corp
Original Assignee
Nippon Telegraph and Telephone Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nippon Telegraph and Telephone Corp filed Critical Nippon Telegraph and Telephone Corp
Priority to JP2001269751A priority Critical patent/JP3716918B2/en
Publication of JP2003078988A publication Critical patent/JP2003078988A/en
Application granted granted Critical
Publication of JP3716918B2 publication Critical patent/JP3716918B2/en
Anticipated expiration legal-status Critical
Expired - Fee Related legal-status Critical Current

Links

Landscapes

  • Circuit For Audible Band Transducer (AREA)

Abstract

PROBLEM TO BE SOLVED: To provide a sound pickup device that separates one sound source signal from a plurality of acoustic signals of sound sources with high S/N. SOLUTION: Two channel sound signals from a microphone are divided into a plurality of frequency bands by each frame, a level or a phase is calculated by each channel and each frequency band, and the levels and the phases from the past to the present frames are weighted-summed. An inter-channel difference of the level or phase subjected to weighted summation is calculated and to which sound source a corresponding frequency band component belongs on the basis of the inter-channel level or phase difference. Frequency band component signals by each channel are synthesized astride the frequency bands to obtain a sound source signal on the basis of a discrimination signal.

Description

【発明の詳細な説明】Detailed Description of the Invention

【0001】[0001]

【発明の属する技術分野】本発明は、空間に複数の音源
が異なる位置に配置されている場合に、少なくとも2本
以上のマイクロホンを用いて、空間を複数のゾーンに分
割し、目的とするゾーンにある音源からの音を他のゾー
ンの音源とは独立に収音する装置、方法及びこの方法を
コンピュータに実行させるプログラム、プログラム記録
媒体に関する。
BACKGROUND OF THE INVENTION 1. Field of the Invention The present invention divides a space into a plurality of zones by using at least two or more microphones when a plurality of sound sources are arranged in different positions in the space, and a desired zone is obtained. The present invention relates to an apparatus, a method, a program for causing a computer to execute the method, a program recording medium, and a method for collecting the sound from a sound source independent of other zones.

【0002】[0002]

【従来の技術】従来のゾーン分離収音技術には、例え
ば、音が持ついくつかの周波数成分の和として表現され
る特徴を利用したものがある。すなわち、複数の音(複
数の音源)が同時に鳴っている場合、互いに離して設け
られた複数のマイクロホンにより収音し、各マイクロホ
ンの各出力チャネル信号を、フレーム毎に複数の周波数
帯域に分割し、その各帯域には主として1つの音源信号
成分のみ存在するようにし、これら分割された各出力チ
ャネル信号の各同一帯域毎に、複数のマイクロホンの位
置に起因して変化する、マイクロホンに到達する音響信
号のパラメータ、すなわちレベル(パワー)、到達時間
の値の差を帯域別チャネル間パラメータ値差として検出
し、各帯域の帯域別チャネル間パラメータ差に基づき、
その帯域の帯域分割された各出力チャネル信号の何れが
何れの音源から入力された信号であるかを判定し、この
音源信号判定に基づき、帯域分割された各出力チャネル
信号の内、同一音源から入力された信号を少なくとも1
つ選択し、選択された複数の帯域信号を音源信号として
合成する音源分離方法が提案されている(参考:特開平
10−313497号公報(特願平09−252312
号)「音源分離方法、装置および記録媒体」)。
2. Description of the Related Art A conventional zone-separated sound collecting technique utilizes, for example, a feature expressed as a sum of several frequency components of a sound. In other words, when multiple sounds (multiple sound sources) are playing at the same time, the sound is picked up by multiple microphones that are provided separately from each other, and each output channel signal of each microphone is divided into multiple frequency bands for each frame. , The sound reaching the microphones, which is caused by the positions of a plurality of microphones, for each of the same bands of the respective divided output channel signals so that only one sound source signal component is present in each of the bands. The signal parameter, that is, the level (power), the difference in the value of the arrival time is detected as the inter-band parameter value difference between bands, and based on the inter-band parameter value difference between each band,
It is determined which of the sound sources is input from which sound source of each band-divided output channel signal of that band, and based on this sound source signal judgment, from the same sound source of each band-divided output channel signal. At least 1 input signal
A sound source separation method has been proposed in which one of the two band signals is selected and synthesized as a sound source signal (reference: Japanese Patent Application Laid-Open No. 10-313497 (Japanese Patent Application No. 09-252312).
No.) "Sound source separation method, device and recording medium").

【0003】[0003]

【発明が解決しようとする課題】ところが、従来の技術
では、部屋の残響時間が長くなるにつれ、チャネル間の
到達レベル差や到達位相差(時間差)の算出誤差が生
じ、その結果、異なるゾーンからの音が混じったり、目
的とするゾーンで発せられている音が劣化するという欠
点があった。
However, in the conventional technique, as the reverberation time of the room becomes longer, an error in calculating the arrival level difference or the arrival phase difference (time difference) between channels occurs, and as a result, the difference from different zones occurs. There is a drawback that the sound of the sound is mixed and the sound emitted in the target zone is deteriorated.

【0004】[0004]

【課題を解決するための手段】上記課題を解決するため
に、 本発明は、従来、1フレームの情報を元に算出さ
れていたチャネル間の到達位相差や到達レベル差を、複
数フレームに渡り加重平均することにより、到達位相差
や到達レベル差の算出誤差を減らし、目的とするゾーン
(音源)の音のみを従来に比べて高S/Nで抽出するこ
とを特徴とする。
SUMMARY OF THE INVENTION In order to solve the above problems, the present invention provides the arrival phase difference and the arrival level difference between channels, which have been calculated based on the information of one frame, over a plurality of frames. By weighted averaging, the calculation error of the arrival phase difference and the arrival level difference is reduced, and only the sound of the target zone (sound source) is extracted with a higher S / N than the conventional one.

【0005】[0005]

【発明の実施の形態】図1に、本発明の実施例である収
音装置の構成を示す。収音手段はそれぞれマイクロホン
1,22で構成し、音源11,12からの音響信号s1(n),s2
(n)(n:時間)を収音して電気信号(チャネル信号)x1
(n),x2(n)に変換する。3は帯域分割手段であり、収音
手段からの信号をフレーム毎に周波数帯域(帯域分解能
が約10〜20Hz)に分割する。その分割した各周波数
帯域信号X1,X2を式(1)、(2)で表す。帯域分割の
手段として例えば、フーリエ変換やウォーブレット変換
があげられる。
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS FIG. 1 shows the configuration of a sound collecting device which is an embodiment of the present invention. Each sound collecting means is constituted by the microphone 2 1, 2 2, the acoustic signal s 1 from the sound source 1 1, 1 2 (n), s 2
(n) (n: time) is picked up and electrical signal (channel signal) x 1
Convert to (n), x 2 (n). A band dividing unit 3 divides the signal from the sound collecting unit into frequency bands (band resolution is about 10 to 20 Hz) for each frame. The divided frequency band signals X 1 and X 2 are represented by equations (1) and (2). Examples of means for band division include Fourier transform and Warblet transform.

【0006】[0006]

【数1】 ここで、ω(2πf)は角速度を表し、lは信号分析長
(フーリエ変換の場合はフレーム長:約20〜40mse
c)のインデックスである。
[Equation 1] Here, ω (2πf) represents an angular velocity, l is a signal analysis length (in the case of Fourier transform, frame length: about 20 to 40 mse).
It is the index of c).

【0007】4のパラメータ値差検出用レベル加重平均
手段においては、X1 ,X2の信号レベルに対し、式(3)
で示す加重平均を行う。
In the parameter-weighted difference detecting level weighted averaging means of 4, the equation (3) is applied to the signal levels of X 1 and X 2.
The weighted average shown in is performed.

【数2】 ここで、0<α≦1とする。また、Lは加重平均に用い
るフレームの個数とする。 i はチャネルのインデック
スである。5のパラメータ値差検出用位相加重平均手段
においては、X1 ,X2の位相に対し、式(4)で示す加重
平均を行う。
[Equation 2] Here, 0 <α ≦ 1. L is the number of frames used for weighted averaging. i is the channel index. In the parameter value difference detecting phase weighted averaging means 5, the weighted average represented by the equation (4) is performed on the phases X 1 and X 2 .

【数3】 ここで、0<β≦1とする。また、Mは加重平均に用い
るフレームの個数とする。
[Equation 3] Here, 0 <β ≦ 1. Further, M is the number of frames used for the weighted average.

【0008】6の信号合成用レベル加重平均手段におい
ては、X1 ,X2の信号レベルに対し、式(5)で示す加重
平均を行う。
In the signal-combining level weighted averaging means 6, the weighted average represented by the equation (5) is applied to the signal levels of X 1 and X 2 .

【数4】 ここで、0<γ≦1とする。また、Nは加重平均に用い
るフレームの個数とする。
[Equation 4] Here, 0 <γ ≦ 1. N is the number of frames used for the weighted average.

【0009】7の信号合成用位相加重平均手段において
は、X1 ,X2の信号レベルに対し、式(6)で示す加重平
均を行う。
In the signal-combining phase weighted averaging means 7, the weighted average represented by the equation (6) is applied to the signal levels of X 1 and X 2 .

【数5】 ここで、0<δ≦1とする。また、Oは加重平均に用い
るフレームの個数とする。パラメータ値差検出用レベ
ル、位相加重平均手段4、5、信号合成用レベル加重平
均手段6,7を設けることにより、α,β,γ,δ及びL,
M,N,Oをそれぞれ異なる値に設定することができる。
[Equation 5] Here, 0 <δ ≦ 1. O is the number of frames used for the weighted average. By providing the parameter value difference detection level, the phase weighted averaging means 4 and 5, and the signal synthesizing level weighted averaging means 6 and 7, α, β, γ, δ and L,
M, N, and O can be set to different values.

【0010】8のパラメータ値差検出手段においては、
パラメータ値差検出用レベル、位相加重平均手段4、5
で加重平均されたレベルまたは位相を用いて、チャネル
間レベル差(ΔLev)、チャネル間位相差(Δarg)を算
出する。これらはそれぞれ、式(7)、(8)で算出さ
れる。
In the parameter value difference detecting means of 8,
Parameter value difference detection level, phase weighted averaging means 4, 5
By using the level or phase weighted and averaged in, the inter-channel level difference (ΔLev) and the inter-channel phase difference (Δarg) are calculated. These are calculated by equations (7) and (8), respectively.

【数6】 [Equation 6]

【0011】9の信号判定手段においては、パラメータ
値差手段8で算出されたチャネル間レベル差またはチャ
ネル間位相差に基づき、|V1(ω,l)|,arg(Y1(ω,l)),|
V2(ω,l)|,arg(Y2(ω,l))に乗算する重み値Wei1
(ω),Wei2(ω)を決定する。例えばチャネル間レベル差
|W1(ω,l)|/|W2(ω,l)|が、ある1より大きな値τよ
り大きい場合、|V1(ω,l)|,arg(Y1(ω,l))には1を乗
算し、|V2(ω,l)|,arg(Y2(ω,l))には0または1より
小さな値a(ω)を乗算するように決定する。すなわち、
あるωにおいてチャネル1の方がレベルが大きい場合に
はゾーン1(チャネル1で代表されるゾーン)に音源が
あると判定する。これに対し、チャネル2の方がレベル
が大きい場合にはゾーン2に音源があると判定する。ま
た、例えば△argが正の値となる場合、|V1(ω,l)|,arg
(Y1(ω,l))には1を乗算し、|V2(ω,l)|,arg(Y2(ω,
l))には0または1より小さな値a(ω)を乗算するよう
に決定する。すなわちチャネル2の方が遅れるのでゾー
ン1に音源があると判定する。但し、ここではarg(X
1(ω,l))等は周波数領域での係数arg(X1(ω,l))等の
複素平面上での位相角の負値である。もし、これを正値
と定義されるなら△argが正の値にある場合にはチャネ
ル1の方が遅れるのでゾーン2と判定する。
In the signal determination means 9 of the above, based on the inter-channel level difference or inter-channel phase difference calculated by the parameter value difference means 8, | V 1 (ω, l) |, arg (Y 1 (ω, l )), |
Weight value Wei1 for multiplying V 2 (ω, l) |, arg (Y 2 (ω, l))
(ω), Wei2 (ω) is determined. Level difference between channels
| W 1 (ω, l) | / | W 2 (ω, l) | is greater than a certain value τ greater than 1 , | V 1 (ω, l) |, arg (Y 1 (ω, l) ) Is multiplied by 1, and | V 2 (ω, l) |, arg (Y 2 (ω, l)) is multiplied by 0 or a value smaller than 1 a (ω). That is,
When the level of channel 1 is higher at a certain ω, it is determined that there is a sound source in zone 1 (zone represented by channel 1). On the other hand, when channel 2 has a higher level, it is determined that there is a sound source in zone 2. Further, for example, when Δarg has a positive value, | V 1 (ω, l) |, arg
(Y 1 (ω, l)) is multiplied by 1 and | V 2 (ω, l) |, arg (Y 2 (ω, l)
l)) is determined to be multiplied by a value a (ω) smaller than 0 or 1. That is, since channel 2 is delayed, it is determined that there is a sound source in zone 1. However, arg (X
1 (ω, l)) and the like are negative values of the phase angle on the complex plane of the coefficient arg (X 1 (ω, l)) and the like in the frequency domain. If this is defined as a positive value, if Δarg has a positive value, channel 1 is delayed and zone 2 is determined.

【0012】10の信号選択手段においては、信号判定
手段9で決定された重み値Wei1(ω)およびWei2(ω)
を、信号合成用レベル、位相加重平均手段6、7から出
力されるレベル及び位相信号に対して乗算する。この場
合、簡略化するために信号合成用レベル及び位相の加重
平均を省略してもよい。
In the signal selection means of 10, the weight values Wei1 (ω) and Wei2 (ω) determined by the signal determination means 9 are used.
Is multiplied by the level for signal synthesis, the level and the phase signal output from the phase weighted averaging means 6, 7. In this case, the weighted average of the signal combining level and the phase may be omitted for simplification.

【0013】11の信号合成手段においては、重み値We
i1(ω),Wei2(ω)が乗算された、信号のレベル|V
1(ω,l)|,|V2(ω,l)|および位相arg(Y1(ω,l)),arg
(Y2(ω,l))を元に、信号を周波数領域から時間領域に
変換することで、各音源からの音s1^(n),s2^(n)を高い
S/Nで抽出する。
In the signal synthesis means of 11, the weight value We
Signal level | V multiplied by i1 (ω) and Wei2 (ω)
1 (ω, l) |, | V 2 (ω, l) | and the phase arg (Y 1 (ω, l)), arg
By converting the signal from the frequency domain to the time domain based on (Y 2 (ω, l)), the sound s 1 ^ (n), s 2 ^ (n) from each sound source can be converted with high S / N. Extract.

【0014】図2を参照して本発明の収音装置の処理を
説明する。互いに離して配置された複数のマイクロホン
からの各出力チャネル信号を入力し、フレーム毎に複数
の周波数帯域に分割する(s1)。複数の周波数帯域に
分割された各チャネル信号X1,X2の各同一帯域毎に、フ
レーム毎に算出された信号のレベル、位相を過去から複
数フレームに渡り保持し、その信号レベル、位相のパラ
メータ値差検出用の加重平均|W1(ω,l)|,|W2(ω,l)|,
arg(U1(ω,l)),arg(U2(ω,l))を算出する(s2)。
また、複数の周波数帯域に分割された各出力チャネル信
号の各同一帯域毎に、フレーム毎に算出された信号のレ
ベル、位相を複数フレームに渡り保持し、その信号レベ
ル、位相の信号合成用の加重平均|V1(ω,l)|,|V2(ω,
l)|,arg(Y1(ω,l)),arg(Y2(ω,l))を算出する(s
3)。複数のマイクロホンの位置に起因して変化する、
マイクロホンに到達する音響信号のパラメータの値の差
としてs2で算出したチャネル間レベル差△Lev、チャ
ネル間位相差△argを検出する(s4)。各帯域のチャ
ネル間パラメータ値差に基づき、その帯域の信号合成用
レベル、位相加重平均した出力信号のうち、何れが何れ
の音源から入力された信号であるかを判定する(すなわ
ち乗算する重み値Wei1(ω),Wei2(ω)を決定する)
(s5)。s5の判定に基づきs3で算出した信号合成
用レベル、位相加重平均信号に重み値を乗算すること
で、同一音源から入力された信号を少なくとも1つ抽出
する(s6)。同一音源からの信号として選択された複
数の帯域信号を合成し音源信号s1^(n),s2^(n)として出
力する(s7)。
The processing of the sound collecting device of the present invention will be described with reference to FIG. Each output channel signal from a plurality of microphones arranged apart from each other is input and divided into a plurality of frequency bands for each frame (s1). For each same band of each channel signal X 1 and X 2 divided into a plurality of frequency bands, the level and phase of the signal calculated for each frame are held over a plurality of frames from the past, and the signal level and phase Weighted average for detecting parameter value difference | W 1 (ω, l) |, | W 2 (ω, l) |,
arg (U 1 (ω, l)), arg (U 2 (ω, l)) are calculated (s2).
Also, for each same band of each output channel signal divided into a plurality of frequency bands, the level and phase of the signal calculated for each frame are held over a plurality of frames, and the signal level and phase for signal combining are held. Weighted average | V 1 (ω, l) |, | V 2 (ω,
l) |, arg (Y 1 (ω, l)), arg (Y 2 (ω, l)) are calculated (s
3). Changes due to the position of multiple microphones,
The inter-channel level difference ΔLev and the inter-channel phase difference Δarg calculated in s2 are detected as the differences in the parameter values of the acoustic signals reaching the microphone (s4). Based on the parameter value difference between channels in each band, it is determined which of the sound source is the signal input from the signal synthesis level and the phase weighted average output signal of that band (that is, the weight value to be multiplied). Wei1 (ω) and Wei2 (ω) are determined)
(S5). At least one signal input from the same sound source is extracted by multiplying the signal combining level and the phase weighted average signal calculated in s3 based on the determination in s5 by a weight value (s6). A plurality of band signals selected as signals from the same sound source are combined and output as sound source signals s 1 ^ (n) and s 2 ^ (n) (s 7 ).

【0015】また、本発明の収音装置は、CPUやメモリ
等を有するコンピュータと、アクセス主体となるユーザ
が利用する利用者端末と、記録媒体から構成することが
できる。記録媒体はCD-ROM、磁気ディスク装置、半導体
メモリ等の機械読み取りが可能な記録媒体であり、ここ
に記録された制御用プログラムは、コンピュータに読み
取られ、コンピュータの動作を制御しコンピュータ上に
前述した実施の形態における各構成要素、すなわち、帯
域分割手段、パラメータ値差検出用レベル、位相加重平
均手段、信号合成用レベル、位相加重平均手段、パラメ
ータ値差検出手段、信号判定手段、信号選択手段、信号
合成手段等を実現する。
The sound collecting device of the present invention can be composed of a computer having a CPU, a memory, etc., a user terminal used by a user who is an access subject, and a recording medium. The recording medium is a machine-readable recording medium such as a CD-ROM, a magnetic disk device, and a semiconductor memory. The control program recorded here is read by a computer, controls the operation of the computer, and operates on the computer. Each component in the embodiment, that is, band dividing means, parameter value difference detecting level, phase weighted averaging means, signal synthesizing level, phase weighted averaging means, parameter value difference detecting means, signal determining means, signal selecting means , Signal synthesizing means, etc. are realized.

【0016】[0016]

【発明の効果】本発明は、チャネル間の到達時間差及び
到達レベル差を加重平均することにより、これら値の算
出誤差を減らし、従来の方法に比べて高いS/Nで目的
とするゾーン(音源)の音を抽出することを可能とす
る。
The present invention reduces the calculation error of these values by weighted averaging the arrival time difference and the arrival level difference between channels, and achieves a target zone (sound source) with a high S / N compared to the conventional method. ) Sound can be extracted.

【図面の簡単な説明】[Brief description of drawings]

【図1】本発明の実施例である収音装置の構成図。FIG. 1 is a configuration diagram of a sound collecting device that is an embodiment of the present invention.

【図2】本発明の実施例である収音装置の処理を説明す
るための図。
FIG. 2 is a diagram for explaining processing of the sound collecting device that is the embodiment of the present invention.

【符号の説明】[Explanation of symbols]

1・・・音源、2・・・マイクロホン、3・・・帯域分
割手段、4・・・パラメータ値差検出用レベル加重平均
手段、5・・・パラメータ値差検出用位相加重平均手
段、6・・・信号合成用レベル加重平均手段、7・・・
信号合成用位相加重平均手段、8・・・パラメータ値差
検出手段、9・・・信号判定手段、10・・・信号選択
手段、15・・・信号合成手段
DESCRIPTION OF SYMBOLS 1 ... Sound source, 2 ... Microphone, 3 ... Band splitting means, 4 ... Parameter value difference detecting level weighted averaging means, 5 ... Parameter value difference detecting phase weighted averaging means, 6 ... ..Level-weighted averaging means for signal synthesis, 7 ...
Phase-weighted average means for signal synthesis, 8 ... Parameter value difference detection means, 9 ... Signal determination means, 10 ... Signal selection means, 15 ... Signal synthesis means

フロントページの続き (51)Int.Cl.7 識別記号 FI テーマコート゛(参考) H04R 3/04 Continuation of front page (51) Int.Cl. 7 Identification code FI theme code (reference) H04R 3/04

Claims (8)

【特許請求の範囲】[Claims] 【請求項1】複数の音源から少なくとも1つの音源を分
離する収音装置において、 少なくとも2チャネルの信号をフレーム毎に複数の周波
数帯域に分割する帯域分割手段と、 各チャネル及び周波数帯域毎にレベル又は位相を算出
し、過去から現在のフレームにわたって加重平均するパ
ラメータ値差検出用加重平均手段と、 前記加重平均されたレベル又は位相のチャネル間差を算
出するチャネル間パラメータ差算出手段と、 前記チャネル間パラメータ差に基づき、対応する周波数
帯域成分がいずれの音源に属するかを判定する信号判定
手段と、 信号判定手段の同一音源からの信号として判定された周
波数帯域成分信号を周波数帯域を跨いで音源信号を合成
する音源合成手段とを有することを特徴とする収音装
置。
1. A sound collecting device for separating at least one sound source from a plurality of sound sources, a band dividing means for dividing a signal of at least two channels into a plurality of frequency bands for each frame, and a level for each channel and each frequency band. Alternatively, a parameter value difference detection weighted averaging unit that calculates a phase and performs a weighted average over the past to the current frame, an inter-channel parameter difference calculation unit that calculates the channel-to-channel difference of the weighted averaged level or phase, and the channel A signal determination unit that determines which sound source the corresponding frequency band component belongs to based on the inter-parameter difference, and a frequency band component signal that is determined as a signal from the same sound source of the signal determination unit A sound pickup device having a sound source synthesizing means for synthesizing signals.
【請求項2】請求項1に記載の収音装置において、 各チャネル及び周波数帯域毎にレベル及び位相を算出
し、過去から現在のフレームにわたって加重平均して周
波数帯域成分信号とし、音源合成手段に出力する信号合
成用加重平均手段を有することを特徴とする収音装置。
2. The sound collecting device according to claim 1, wherein a level and a phase are calculated for each channel and a frequency band, and a weighted average is applied over a past frame to a present frame to obtain a frequency band component signal, and the sound source synthesizing means A sound collecting device having a weighted averaging means for outputting signals.
【請求項3】複数の音源から少なくとも1つの音源を分
離する収音方法において、 少なくとも2チャネルの信号をフレーム毎に複数の周波
数帯域に分割し、各チャネル及び周波数帯域毎にレベル
又は位相を算出し、 前記レベル及び位相を過去から現在のフレームにわたっ
て加重平均し、 前記加重平均されたレベル又は位相のチャネル間差を算
出し、 前記加重平均されたレベル又は位相のチャネル間差に基
づき、対応する周波数帯域成分がいずれの音源に属する
か判定し、 同一音源からの信号として判定された周波数帯域成分信
号を周波数帯域を跨いで音源信号を合成することを特徴
とする収音方法。
3. A sound pickup method for separating at least one sound source from a plurality of sound sources, wherein a signal of at least two channels is divided into a plurality of frequency bands for each frame, and a level or phase is calculated for each channel and each frequency band. Then, the level and the phase are weighted averaged from the past to the current frame, the difference between the channels of the weighted average of the level or the phase is calculated, and the difference between the channels of the weighted average of the level or the phase is calculated. A sound collecting method characterized by determining which sound source a frequency band component belongs to, and synthesizing a sound source signal across frequency bands from the frequency band component signals determined as signals from the same sound source.
【請求項4】請求項3に記載の収音方法において、 前記周波数帯域成分信号は、各チャネル及び周波数帯域
毎にレベル及び位相を過去から現在のフレームにわたっ
て加重平均した信号であることを特徴とする収音方法。
4. The sound collecting method according to claim 3, wherein the frequency band component signal is a signal obtained by weighting and averaging levels and phases for each channel and each frequency band over the past to the present frame. How to collect sound.
【請求項5】少なくとも2チャネルの信号をフレーム毎
に複数の周波数帯域に分割する処理と、 各チャネル及び周波数帯域毎にレベル又は位相を算出
し、過去から現在のフレームにわたって加重平均する処
理と、 前記加重平均されたレベル又は位相のチャネル間差を算
出する処理と、 前記加重平均されたレベル又は位相のチャネル間差に基
づき、対応する周波数帯域成分がいずれの音源に属する
か判定する処理と、 同一音源からの信号として判定された周波数帯域成分信
号を周波数帯域を跨いで音源信号を合成する処理とをコ
ンピュータに実行させるための複数の音源から少なくと
も1つの音源を分離する収音方法のプログラム。
5. A process of dividing a signal of at least two channels into a plurality of frequency bands for each frame, a process of calculating a level or a phase for each channel and a frequency band, and performing a weighted average over the past to the present frame. A process of calculating the difference between the channels of the weighted average level or phase, and a process of determining which sound source the corresponding frequency band component belongs to, based on the difference between the channels of the weighted average level or phase, A program of a sound collection method for separating at least one sound source from a plurality of sound sources for causing a computer to perform a process of synthesizing a sound source signal of a frequency band component signal determined as a signal from the same sound source across frequency bands.
【請求項6】請求項5に記載の収音方法のプログラムに
おいて、 各チャネル及び周波数帯域毎にレベル及び位相を算出
し、過去から現在のフレームにわたって加重平均して周
波数帯域成分信号とする処理を有することを特徴とする
収音方法のプログラム。
6. The program of the sound collecting method according to claim 5, wherein a level and a phase are calculated for each channel and a frequency band, and a weighted average is performed over the past to the current frame to obtain a frequency band component signal. A program for a sound collection method having.
【請求項7】少なくとも2チャネルの信号をフレーム毎
に複数の周波数帯域に分割する処理と、 各チャネル及び周波数帯域毎にレベル又は位相を算出
し、過去から現在のフレームにわたって加重平均する処
理と、 前記加重平均されたレベル又は位相のチャネル間差を算
出する処理と、 前記加重平均されたレベル又は位相のチャネル間差に基
づき、対応する周波数帯域成分がいずれの音源に属する
か判定する処理と、 同一音源からの信号として判定された周波数帯域成分信
号を周波数帯域を跨いで音源信号を合成する処理とをコ
ンピュータに実行させるための複数の音源から少なくと
も1つの音源を分離する収音方法のプログラムを記録し
たコンピュータ読み取り可能な記録媒体。
7. A process of dividing a signal of at least two channels into a plurality of frequency bands for each frame, a process of calculating a level or a phase for each channel and a frequency band, and performing a weighted average over the past to the present frame. A process of calculating the difference between the channels of the weighted average level or phase, and a process of determining which sound source the corresponding frequency band component belongs to, based on the difference between the channels of the weighted average level or phase, A sound collection method program for separating at least one sound source from a plurality of sound sources for causing a computer to execute processing of synthesizing sound source signals across frequency bands of frequency band component signals determined as signals from the same sound source. The recorded computer-readable recording medium.
【請求項8】請求項7に記載の記録媒体において、 各チャネル及び周波数帯域毎にレベル及び位相を算出
し、過去から現在のフレームにわたって加重平均して周
波数帯域成分信号とする処理を有する記録媒体。
8. The recording medium according to claim 7, further comprising a process of calculating a level and a phase for each channel and a frequency band, and performing a weighted average over a past frame to a present frame to obtain a frequency band component signal. .
JP2001269751A 2001-09-06 2001-09-06 Sound collection device, method and program, and recording medium Expired - Fee Related JP3716918B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
JP2001269751A JP3716918B2 (en) 2001-09-06 2001-09-06 Sound collection device, method and program, and recording medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
JP2001269751A JP3716918B2 (en) 2001-09-06 2001-09-06 Sound collection device, method and program, and recording medium

Publications (2)

Publication Number Publication Date
JP2003078988A true JP2003078988A (en) 2003-03-14
JP3716918B2 JP3716918B2 (en) 2005-11-16

Family

ID=19095514

Family Applications (1)

Application Number Title Priority Date Filing Date
JP2001269751A Expired - Fee Related JP3716918B2 (en) 2001-09-06 2001-09-06 Sound collection device, method and program, and recording medium

Country Status (1)

Country Link
JP (1) JP3716918B2 (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2007010897A (en) * 2005-06-29 2007-01-18 Toshiba Corp Sound signal processing method, device, and program
JP2007074665A (en) * 2005-09-09 2007-03-22 Nippon Telegr & Teleph Corp <Ntt> Collected sound reproducing apparatus
JP2007306373A (en) * 2006-05-12 2007-11-22 Nippon Telegr & Teleph Corp <Ntt> Apparatus, method, and program for signal separation, and recording medium
JPWO2006090589A1 (en) * 2005-02-25 2008-07-24 パイオニア株式会社 Sound separation device, sound separation method, sound separation program, and computer-readable recording medium
JP2009010992A (en) * 2008-09-01 2009-01-15 Sony Corp Audio signal processing apparatus, audio signal processing method, and program
JP2009010996A (en) * 2008-09-11 2009-01-15 Sony Corp Sound signal processor and processing method
JP2009506363A (en) * 2005-08-26 2009-02-12 ステップ・コミュニケーションズ・コーポレーション Method and apparatus for adapting to device and / or signal mismatch in a sensor array
JP2009147654A (en) * 2007-12-13 2009-07-02 Sony Corp Sound processor, sound processing system, and sound processing program
JP2012507049A (en) * 2008-10-24 2012-03-22 クゥアルコム・インコーポレイテッド System, method, apparatus and computer readable medium for coherence detection
US8155927B2 (en) 2005-08-26 2012-04-10 Dolby Laboratories Licensing Corporation Method and apparatus for improving noise discrimination in multiple sensor pairs
US8620672B2 (en) 2009-06-09 2013-12-31 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for phase-based processing of multichannel signal
US8886499B2 (en) 2011-12-27 2014-11-11 Fujitsu Limited Voice processing apparatus and voice processing method

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4767247B2 (en) * 2005-02-25 2011-09-07 パイオニア株式会社 Sound separation device, sound separation method, sound separation program, and computer-readable recording medium
JPWO2006090589A1 (en) * 2005-02-25 2008-07-24 パイオニア株式会社 Sound separation device, sound separation method, sound separation program, and computer-readable recording medium
JP2007010897A (en) * 2005-06-29 2007-01-18 Toshiba Corp Sound signal processing method, device, and program
US7995767B2 (en) 2005-06-29 2011-08-09 Kabushiki Kaisha Toshiba Sound signal processing method and apparatus
US8155927B2 (en) 2005-08-26 2012-04-10 Dolby Laboratories Licensing Corporation Method and apparatus for improving noise discrimination in multiple sensor pairs
JP2009506363A (en) * 2005-08-26 2009-02-12 ステップ・コミュニケーションズ・コーポレーション Method and apparatus for adapting to device and / or signal mismatch in a sensor array
JP2007074665A (en) * 2005-09-09 2007-03-22 Nippon Telegr & Teleph Corp <Ntt> Collected sound reproducing apparatus
JP4616736B2 (en) * 2005-09-09 2011-01-19 日本電信電話株式会社 Sound collection and playback device
JP2007306373A (en) * 2006-05-12 2007-11-22 Nippon Telegr & Teleph Corp <Ntt> Apparatus, method, and program for signal separation, and recording medium
JP4676920B2 (en) * 2006-05-12 2011-04-27 日本電信電話株式会社 Signal separation device, signal separation method, signal separation program, and recording medium
JP2009147654A (en) * 2007-12-13 2009-07-02 Sony Corp Sound processor, sound processing system, and sound processing program
JP2009010992A (en) * 2008-09-01 2009-01-15 Sony Corp Audio signal processing apparatus, audio signal processing method, and program
JP2009010996A (en) * 2008-09-11 2009-01-15 Sony Corp Sound signal processor and processing method
JP2012507049A (en) * 2008-10-24 2012-03-22 クゥアルコム・インコーポレイテッド System, method, apparatus and computer readable medium for coherence detection
US8724829B2 (en) 2008-10-24 2014-05-13 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for coherence detection
US8620672B2 (en) 2009-06-09 2013-12-31 Qualcomm Incorporated Systems, methods, apparatus, and computer-readable media for phase-based processing of multichannel signal
US8886499B2 (en) 2011-12-27 2014-11-11 Fujitsu Limited Voice processing apparatus and voice processing method

Also Published As

Publication number Publication date
JP3716918B2 (en) 2005-11-16

Similar Documents

Publication Publication Date Title
JP5229053B2 (en) Signal processing apparatus, signal processing method, and program
CN100356445C (en) Method and apparatus for separating sound-source signal and method and device for detecting pitch
US7095865B2 (en) Audio amplifier unit
JP4449987B2 (en) Audio processing apparatus, audio processing method and program
US9154895B2 (en) Apparatus of generating multi-channel sound signal
JP3716918B2 (en) Sound collection device, method and program, and recording medium
US20130021502A1 (en) Sound corrector, sound recording device, sound reproducing device, and sound correcting method
JP2010187363A (en) Acoustic signal processing apparatus and reproducing device
JPH0997091A (en) Method for pitch change of prerecorded background music and karaoke system
JP4670682B2 (en) Audio apparatus and directional sound generation method
WO2006057131A1 (en) Sound reproducing device and sound reproduction system
JP2003270034A (en) Sound information analyzing method, apparatus, program, and recording medium
JP3033061B2 (en) Voice noise separation device
JP2005512434A (en) Circuit and method for enhancing a stereo signal
JP2005266797A (en) Method and apparatus for separating sound-source signal and method and device for detecting pitch
KR20060034637A (en) Method and device for removing known acoustic signal
JPH07240990A (en) Microphone device
JP2016163135A (en) Sound collection device, program and method
JP2004325127A (en) Sound source detection method, sound source separation method, and apparatus for executing them
JP3588576B2 (en) Sound pickup device and sound pickup method
JP4249697B2 (en) Sound source separation learning method, apparatus, program, sound source separation method, apparatus, program, recording medium
EP3513573B1 (en) A method, apparatus and computer program for processing audio signals
JP3787103B2 (en) Speech processing apparatus, speech processing method, speech processing program
JP2005062096A (en) Detection method of speaker position, system, program and record medium
JP2009282536A (en) Method and device for removing known acoustic signal

Legal Events

Date Code Title Description
A977 Report on retrieval

Free format text: JAPANESE INTERMEDIATE CODE: A971007

Effective date: 20050728

TRDD Decision of grant or rejection written
A01 Written decision to grant a patent or to grant a registration (utility model)

Free format text: JAPANESE INTERMEDIATE CODE: A01

Effective date: 20050802

RD03 Notification of appointment of power of attorney

Free format text: JAPANESE INTERMEDIATE CODE: A7423

Effective date: 20050823

A61 First payment of annual fees (during grant procedure)

Free format text: JAPANESE INTERMEDIATE CODE: A61

Effective date: 20050823

R150 Certificate of patent or registration of utility model

Free format text: JAPANESE INTERMEDIATE CODE: R150

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20080909

Year of fee payment: 3

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20090909

Year of fee payment: 4

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20090909

Year of fee payment: 4

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20100909

Year of fee payment: 5

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20100909

Year of fee payment: 5

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20110909

Year of fee payment: 6

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20120909

Year of fee payment: 7

FPAY Renewal fee payment (event date is renewal date of database)

Free format text: PAYMENT UNTIL: 20130909

Year of fee payment: 8

LAPS Cancellation because of no payment of annual fees