CN101604526B - Weight-based system and method for calculating audio frequency attention - Google Patents

Weight-based system and method for calculating audio frequency attention Download PDF

Info

Publication number
CN101604526B
CN101604526B CN2009100630452A CN200910063045A CN101604526B CN 101604526 B CN101604526 B CN 101604526B CN 2009100630452 A CN2009100630452 A CN 2009100630452A CN 200910063045 A CN200910063045 A CN 200910063045A CN 101604526 B CN101604526 B CN 101604526B
Authority
CN
China
Prior art keywords
attention
sound
mrow
weight
interference
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN2009100630452A
Other languages
Chinese (zh)
Other versions
CN101604526A (en
Inventor
胡瑞敏
杭波
马晔
高戈
杨玉红
周成
王晓晨
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Original Assignee
Wuhan University WHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU filed Critical Wuhan University WHU
Priority to CN2009100630452A priority Critical patent/CN101604526B/en
Publication of CN101604526A publication Critical patent/CN101604526A/en
Application granted granted Critical
Publication of CN101604526B publication Critical patent/CN101604526B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Circuit For Audible Band Transducer (AREA)

Abstract

The invention relates to the technical field for processing audio frequency signals, in particular to a weight-based system and a method for calculating audio frequency attention. The system comprises an initialization module, an interesting sound attention calculation module, a switch module, a disturbing sound attention calculation module and an attention fusion module, wherein the attention fusion module calculates the intermediate results of the modules so as to obtain the weight-based attention M of interesting sound and the weight-based attention N of disturbing sound, finally calculates to obtain the fusion attention Ma which is obtained after attention fusion and output the fusion attention. The invention has the advantages that a proper attention judgment threshold value of the audio frequency attention can be easily set under different total types of interesting sound, thereby reducing the difficulty of judging the audio frequency attention and increasing the accuracy of calculating and judging the audio frequency attention.

Description

System and method for calculating audio attention based on weight
Technical Field
The invention relates to the technical field of audio signal processing, in particular to a system and a method for calculating audio attention based on weight.
Background
Attention is the degree of human attention to objects, while audio attention is attention to audio objects and features. Currently, existing audio attention calculation methods include a top-down method and a bottom-up method. The top-down audio attention calculation method is characterized in that whether an object is an attention object is determined by classifying and judging the object according to different attention degrees of people on specific class objects, such as different attention degrees of different types of sounds such as explosion sound and voice in audio; the bottom-up audio attention calculation method is to determine whether an object belongs to an object of interest by analyzing features of the object according to human attention to the features of the object, such as energy and frequency.
The following describes a technical scheme of an audio attention calculation method, which is a common audio attention calculation method in the prior art, in detail.
When a piece of audio contains multiple types of sounds, attention (M) describing the various types of sounds of interest in the piece of audio is first sought, usually based on audio classification1,M2… …) and then further calculates the audio attention M of the audio file at a certain momenta
<math> <mrow> <msub> <mi>M</mi> <mi>a</mi> </msub> <mo>=</mo> <msubsup> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </msubsup> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>,</mo> </mrow> </math> And is <math> <mrow> <msubsup> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </msubsup> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <mo>=</mo> <mn>1</mn> <mo>.</mo> </mrow> </math>
Wherein M isiAttention, λ, to the i-th class of sounds of interest in a segment of audioiCalculating the audio attention M for the ith type interesting sound contained in the audioaThe weight of time, n, is the total number of sound types of interest contained in the piece of audio. And MiThe calculation method is as follows:
Mi=E·Pi
wherein E ═ Eavr/Max_Eavr,EavrMax _ E, the average energy of the i-th class of sounds of interest in the piece of audioavrFor the whole audio file E in which the section of audio is positionedavrIs measured.
And P isi=exp(si-Max_si),siLog-likelihood score (Max _ s) of i-th class of sound of interest for the segment of audioiIs the whole audio file s of the section of audioiIs measured.
If the type of the interesting sound which needs to be included in the audio attention calculation is less, for example, only 2 to 3 types of interesting sounds are included in the audio, the audio attention calculation method can better judge the interesting soundIn other words, when the user focuses on a plurality of sound types, the different types of interesting sounds are given a weight λiWill decline in a reciprocal curve, i.e. as n increases, MaWill be much less than 1, or even close to 0; meanwhile, under different conditions, the number of types of interesting sounds is different, so that when the audio attention is determined by the audio attention calculation method in the prior art, it is inconvenient to select a proper attention determination threshold value for the audio sequence under different conditions, and the determination of the audio attention is complex and inaccurate.
Considering that there are some types of sounds in audio that are not sounds of interest but have some similar features to sounds of interest that need to be included in the audio attention calculation, it is possible that they will be misjudged as sounds of interest, which we call interference sounds. If the effects of these types of sounds are not excluded from the audio attention calculation, the accuracy of the audio attention may be affected.
Disclosure of Invention
The invention aims to provide a weight-based audio attention calculation system and method, which enable the audio attention to be easily set to be a proper attention judgment threshold value under the condition that the total number of interesting sound types is different, and eliminate the influence of interference sound types similar to the interesting sound types in audio when the audio attention is calculated.
In order to achieve the purpose, the invention adopts the following technical scheme:
a weight-based audio attention calculation system, comprising:
an initialization module for setting the total number i of the interested sound types and the weight lambda of each interested soundiSetting the total number j of the types of the interference sounds if the interference sounds to be eliminated exist, and setting the weight value omega of each type of the interference soundsj
Interesting sound attention calculationA module for receiving the total number i of the interested sound types output by the initialization module, detecting the i-type interested sounds and respectively calculating the attention M of the i-th type interested soundi
The switch module is used for receiving the total number j of the interference sound types output by the initialization module and judging whether the total number j of the interference sound types needs to be output to the interference sound attention calculation module;
an interference sound attention degree calculation module for receiving the total number j of the types of the interference sounds output by the switch module, detecting the j types of the interference sounds and respectively calculating the attention degree N of the j type of the interference soundsj
An attention fusion module for receiving the weight value lambda of each interesting sound output from the initialization moduleiAnd the weight omega of each interference soundjReceiving attention M of the i-th type of interesting sound output by the interesting sound attention calculation moduleiReceiving attention degree N of j-th interference sound output by the interference sound attention degree calculation modulej(ii) a And calculating the intermediate result to obtain the attention M based on weight of the interesting sound and the attention N based on weight of the interference sound, and finally calculating to obtain the fusion attention M obtained after attention fusionaAnd outputs the fusion attention degree,
wherein the total number of interesting sound types i is greater than 0 and the total number of disturbing sound types j is greater than or equal to 0.
The interesting sound attention degree calculation module is obtained by adopting any existing audio attention degree calculation method when the attention degree of the interesting sound is calculated.
In the interesting sound attention calculation module, the selection range of the weights of all types of interesting sounds is from 0 to 1, including 0 and 1.
In the disturbing sound attention degree calculation module, the weight of each kind of disturbing sound is selected from 0 to 1, including 0 and 1.
The method for calculating the audio attention based on the weight comprises the following steps:
step 1: inputting an audio signal to be calculated, and initializing the total number i of interested sound types and the total number j of interference sound types;
step 2: detecting i-type interesting sounds in the audio signal and calculating the attention M of various interesting soundsi
And step 3: setting the weight lambda of various interesting sounds in the audio signaliAnd a weight-based audio attention of the sound of interest is calculated, that is,
<math> <mrow> <mi>M</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> λi∈[0,1]
and 4, step 4: judging whether j is 0 or not according to the total number j of the interference sound types initialized in the step 1; if j is 0, the weight-based audio attention N of the disturbing sound is 0, and step 7 is performed; otherwise, executing step 5;
and 5: detecting j interference sounds in the audio signal, and calculating attention degree N of the interference soundsj
Step 6: setting the weight omega of various interference sounds in the audio signaljAnd a weight-based audio attention calculation for each type of disturbing sound, that is,
<math> <mrow> <mi>N</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>0</mn> </mrow> <mi>m</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&omega;</mi> <mi>j</mi> </msub> <msub> <mi>N</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> ωj∈[0,1]
and 7: calculating a fusion attention M according to the audio attention M based on the weight of the interesting sound and the attention N based on the weight of the interference sound obtained in the stepaThat is to say that,
Ma=M(1-N)。
the invention has the following advantages and positive effects:
1) the audio attention is easy to set a proper attention judgment threshold value under the condition that the total number of the interesting sound types is different;
2) excluding the influence of interfering sound types similar to the type of sound of interest in the audio;
3) the judgment difficulty of the audio attention is reduced, and the calculation and judgment accuracy of the audio attention is improved.
Drawings
FIG. 1 is a block diagram of a system for calculating audio attention based on weight according to the present invention.
Fig. 2 is a flowchart of a method for calculating audio attention based on weight according to the present invention.
Fig. 3 is a graph showing the calculation results of the audio attention of 10 sounds of interest in the embodiment of the present invention.
Fig. 4 is a graph showing the calculation result of the audio attention of 10 kinds of interesting sounds and 1 kind of disturbing sounds to be eliminated in the embodiment of the present invention.
Fig. 5 is a graph of the result of audio attention calculation based on weights in the embodiment of the present invention.
Fig. 6 is a diagram of the result of weight-based audio attention calculation without considering sounds to be excluded in the embodiment of the present invention.
Fig. 7 is a diagram of an audio attention calculation result obtained by using a conventional audio attention calculation method according to an embodiment of the present invention.
Wherein,
the method comprises the following steps of 1-initializing module, 2-interesting sound attention calculating module, 3-switching module, 4-interference sound attention calculating module and 5-attention fusing module.
Detailed Description
The invention is further illustrated by the following specific examples in conjunction with the accompanying drawings:
the invention provides a weight-based audio attention calculation system, which specifically adopts the following technical scheme, referring to fig. 1, and comprises an initialization module 1, an interesting sound attention calculation module 2, a switch module 3, an interference sound attention calculation module 4 and an attention fusion module 5.
An initialization module 1 for setting the total number i of the types of the sounds of interest and the weight λ of each sound of interestiSetting the total number j of the types of the interference sounds if the interference sounds to be eliminated exist, and setting the weight value omega of each type of the interference soundsj
An interesting sound attention degree calculation module 2 for receiving the total number i of interesting sound types output from the initialization module 1 and detecting the i types of interesting soundsMeasuring and respectively calculating the attention M of the i-th class of interesting soundi
The switch module 3 is used for receiving the total number j of the interference sound types output by the initialization module 1 and judging whether the total number j of the interference sound types needs to be output to the interference sound attention degree calculation module 4;
an interference sound attention degree calculation module 4, configured to receive the total number j of the types of interference sounds output by the switch module 3, detect j types of interference sounds, and calculate attention degrees N of j types of interference sounds respectivelyj
An attention fusion module 5 for receiving the weight λ of each sound of interest output from the initialization module 1iAnd the weight omega of each interference soundjReceiving attention M from the i-th type of sound of interest outputted from the attention calculation block 2iReceiving attention degree N of j-th interference sound output by interference sound attention degree calculation module 4j(ii) a And calculating the intermediate result to obtain the attention M based on weight of the interesting sound and the attention N based on weight of the interference sound, and finally calculating to obtain the fusion attention M obtained after attention fusionaAnd outputs fusion attention Ma
Under an embodiment of the present invention, it may be set that each type of interesting sound has the same attention degree, each type of interesting sound has the same weight when calculating the attention degree, and λ may be takeni=ωj=1。
The method for calculating the audio attention based on the weight provided by the invention specifically adopts the following technical scheme, referring to fig. 2, and comprises the following steps:
step 1: inputting an audio signal to be calculated, and initializing the total number i of interested sound types and the total number j of interference sound types;
step 2: detecting a sense of class i in the audio signalInteresting sound, and calculating the attention M of various interesting soundsi
And step 3: setting the weight lambda of various interesting sounds in the audio signaliAnd a weight-based audio attention of the sound of interest is calculated, that is,
<math> <mrow> <mi>M</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> λi∈[0,1]
and 4, step 4: judging whether j is 0 or not according to the total number j of the interference sound types initialized in the step 1; if j is 0, the weight-based audio attention N of the disturbing sound is 0, and step 7 is performed; otherwise, executing step 5;
and 5: detecting j interference sounds in the audio signal, and calculating attention degree N of the interference soundsj
Step 6: setting the weight omega of various interference sounds in the audio signaljAnd a weight-based audio attention calculation for each type of disturbing sound, that is,
<math> <mrow> <mi>N</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>0</mn> </mrow> <mi>m</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&omega;</mi> <mi>j</mi> </msub> <msub> <mi>N</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> ωj∈[0,1]
and 7: calculating a fusion attention M according to the audio attention M based on the weight of the interesting sound and the attention N based on the weight of the interference sound obtained in the stepaThat is to say that,
Ma=M(1-N)
one embodiment of the weight-based audio attention calculation method of the present invention is described as follows:
an audio signal includes 11 different sound types, wherein the first 10 sound types are the sound types of interest, the 6 th sound type is speech, the 11 th sound type is the interfering sound type to be excluded, and the sound type is broadcast speech.
Inputting audio signals to be calculated, namely audio signals containing 11 different sound types, and setting the total number i of interesting sound types to be 10 and the total number j of interference sound types to be 1;
calculating the attention of each type of interesting sound in the audio signal, namely the attention of the 10 interesting sounds is M in sequence1,M2,...M10(see FIG. 3);
calculating the attention of each type of sound to be excluded in the audio signal, namely M11(see FIG. 4);
setting the weights of various interesting sounds in the audio signal, namely the weights of 10 interesting sounds are lambda1=λ2...=λ10=1;
Setting the weight of various interference sounds in the audio signal, namely the weight of the 11 th interference sound to be eliminated, namely the broadcast voice, to be omega1=1;
The weight-based audio attention calculation of the sounds of interest is performed on the attentions and weights of the above-mentioned 10 kinds of sounds of interest, that is,
<math> <mrow> <mi>M</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>10</mn> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> λ1=λ2=...=λ10=1
the audio attention calculation based on the weight of the interference sound is performed on the attention and the weight of the above-mentioned 1 kinds of interference sounds, that is,
N=1-(1-ω1N1),ω1=1
calculating fusion attention Ma3I.e. Ma3=M(1-N);
The audio attention based on the weight, which outputs 10 sounds including a gunshot sound, an alarm sound, a voice, a car whistle sound, etc., and excludes the audio signal of the broadcast voice, is Ma3(see FIG. 5);
in the above steps, if the weight of each type of interference sound in the audio signal is set to 0, i.e. the weight of the 11 th type of sound broadcast voice is ω 10, i.e., without considering the influence of the sound to be disturbed on the attention, is available (see fig. 6),
<math> <mrow> <msub> <mi>M</mi> <mrow> <mi>a</mi> <mn>2</mn> </mrow> </msub> <mo>=</mo> <mn>1</mn> <mo>-</mo> <munderover> <mi>&Pi;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>10</mn> </munderover> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> </mrow> </math>
in the above-mentioned steps, if the calculation is performed according to the existing audio attention calculation method described in the background art, that is, the calculation is performed according to the existing audio attention calculation method
<math> <mrow> <msub> <mi>M</mi> <mrow> <mi>a</mi> <mn>1</mn> </mrow> </msub> <mo>=</mo> <msubsup> <mi>&Sigma;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mn>10</mn> </msubsup> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>,</mo> </mrow> </math> And lambdai=1/10=0.1
Then Ma1The resulting audio attention is calculated for the existing audio attention calculation method (see fig. 7).
Comparing fig. 5 and fig. 6, the audio attention calculated by the weight-based audio attention calculation method according to the present invention excludes the influence of the disturbing sound type in the audio, i.e., broadcast voice, on the audio attention calculation.
Comparing fig. 6 and fig. 7, the maximum value of the audio attention value calculated by the weight-based audio attention calculation method according to the present invention is always close to 1, while the audio attention value calculated by the conventional attention calculation method is continuously decreased as the total number of the types of the sounds of interest in the audio increases. Therefore, compared with the existing audio attention calculation method, the audio attention calculation method provided by the invention is more convenient for setting a proper attention judgment threshold value under the condition that the total number of the interested sounds is different, so that the attention calculation and judgment are simpler and more accurate.

Claims (4)

1. A weight-based audio attention calculation system, comprising:
an initialization module (1) for setting the total number i of sound types of interest and the weight λ of each sound of interestiSetting the total number j of the types of the interference sounds if the interference sounds to be eliminated exist, and setting the weight value omega of each type of the interference soundsj
An interesting sound attention degree calculation module (2) for receiving the total number i of interesting sound types output by the initialization module (1) and carrying out I types of interesting soundsDetecting and respectively calculating the attention M of the i-th class of interesting soundi
The switch module (3) is used for receiving the total number j of the interference sound types output by the initialization module (1) and judging whether the total number j of the interference sound types needs to be output to the interference sound attention calculation module (4);
an interference sound attention degree calculation module (4) for receiving the total number j of the types of the interference sounds output by the switch module (3), detecting the j types of the interference sounds and respectively calculating the attention degree N of the j type of the interference soundsj
A focus fusion module (5) for receiving the weight lambda of each interesting sound output from the initialization module (1)iAnd the weight omega of each interference soundjReceiving attention M of the ith type of interesting sound output by the interesting sound attention calculation module (2)iReceiving attention degree N of j-th interference sound output by the interference sound attention degree calculation module (4)j(ii) a Calculating the intermediate result input by the attention fusion module to obtain the attention M of the interested sound based on weight and the attention N of the interference sound based on weight, and finally calculating to obtain the fusion attention M obtained after attention fusionaAnd outputs fusion attention MaWherein:
Ma=M(1-N),
<math> <mrow> <mi>M</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> λi∈[0,1];
<math> <mrow> <mi>N</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>0</mn> </mrow> <mi>m</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&omega;</mi> <mi>j</mi> </msub> <msub> <mi>N</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> ωj∈[0,1](ii) a m and n are positive integers
Wherein the total number of interesting sound types i is greater than 0 and the total number of disturbing sound types j is greater than or equal to 0.
2. The weight-based audio attention computing system of claim 1, wherein:
in the interesting sound attention calculation module, the selection range of the weights of all types of interesting sounds is from 0 to 1, including 0 and 1.
3. The weight-based audio attention computing system of claim 1, wherein:
in the disturbing sound attention degree calculation module, the weight of each kind of disturbing sound is selected from 0 to 1, including 0 and 1.
4. A method for calculating audio attention based on weight is characterized by comprising the following steps:
step 1: inputting an audio signal to be calculated, and initializing the total number i of interested sound types and the total number j of interference sound types;
step 2: detecting i-type interesting sounds in the audio signal and calculating the attention M of various interesting soundsi
And step 3: the weight λ i of each type of sound of interest in the audio signal is set, and the weight-based audio attention of the sound of interest is calculated, i.e.,
<math> <mrow> <mi>M</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>i</mi> <mo>=</mo> <mn>1</mn> </mrow> <mi>n</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&lambda;</mi> <mi>i</mi> </msub> <msub> <mi>M</mi> <mi>i</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> λi∈[0,1]n is a positive integer
And 4, step 4: judging whether j is 0 or not according to the total number j of the interference sound types initialized in the step 1; if j is 0, the weight-based audio attention N of the disturbing sound is 0, and step 7 is performed; otherwise, executing step 5;
and 5: detecting j interference sounds in the audio signal, and calculating attention degree N of the interference soundsj
Step 6: setting the weight omega of various interference sounds in the audio signaljAnd a weight-based audio attention calculation for each type of disturbing sound, that is,
<math> <mrow> <mi>N</mi> <mo>=</mo> <mn>1</mn> <mo>-</mo> <msubsup> <mi>&Pi;</mi> <mrow> <mi>j</mi> <mo>=</mo> <mn>0</mn> </mrow> <mi>m</mi> </msubsup> <mrow> <mo>(</mo> <mn>1</mn> <mo>-</mo> <msub> <mi>&omega;</mi> <mi>j</mi> </msub> <msub> <mi>N</mi> <mi>j</mi> </msub> <mo>)</mo> </mrow> <mo>,</mo> </mrow> </math> ωj∈[0,1]m is a positive integer
And 7: calculating a fusion attention M according to the audio attention M based on the weight of the interesting sound and the attention N based on the weight of the interference sound obtained in the stepaThat is to say that,
Ma=M(1-N)。
CN2009100630452A 2009-07-07 2009-07-07 Weight-based system and method for calculating audio frequency attention Expired - Fee Related CN101604526B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN2009100630452A CN101604526B (en) 2009-07-07 2009-07-07 Weight-based system and method for calculating audio frequency attention

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN2009100630452A CN101604526B (en) 2009-07-07 2009-07-07 Weight-based system and method for calculating audio frequency attention

Publications (2)

Publication Number Publication Date
CN101604526A CN101604526A (en) 2009-12-16
CN101604526B true CN101604526B (en) 2011-11-16

Family

ID=41470232

Family Applications (1)

Application Number Title Priority Date Filing Date
CN2009100630452A Expired - Fee Related CN101604526B (en) 2009-07-07 2009-07-07 Weight-based system and method for calculating audio frequency attention

Country Status (1)

Country Link
CN (1) CN101604526B (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109040778B (en) * 2018-09-12 2021-01-22 武汉轻工大学 Video cover determining method, user equipment, storage medium and device

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1282114A1 (en) * 2001-01-25 2003-02-05 Sony Corporation Data processing apparatus
CN1459093A (en) * 2001-01-25 2003-11-26 索尼公司 Data processing device
CN101419801A (en) * 2008-12-03 2009-04-29 武汉大学 Method for subband measuring correlation sensing characteristic between ears and device thereof

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1282114A1 (en) * 2001-01-25 2003-02-05 Sony Corporation Data processing apparatus
CN1459093A (en) * 2001-01-25 2003-11-26 索尼公司 Data processing device
CN101419801A (en) * 2008-12-03 2009-04-29 武汉大学 Method for subband measuring correlation sensing characteristic between ears and device thereof

Non-Patent Citations (14)

* Cited by examiner, † Cited by third party
Title
.基于用户关注空间与注意力分析的视频精彩摘要与排序.《计算机学报》.2008,
Hu Ruimin
Ruan Ruolin
Ruan Ruolin;Xiao Xuqing.A rate control scheme based on MAD weighted model for H.264/AVC.《2007 3rd International Conference on Wireless Communications, Networking, and Mobile Computing- WiCOM 07》.2007, *
Xiao Xuqing.A rate control scheme based on MAD weighted model for H.264/AVC.《2007 3rd International Conference on Wireless Communications, Networking, and Mobile Computing - WiCOM 07》.2007,
Yang Yuhong
Yang Yuhong;Hu Ruimin;Zhang Yong;Zhang Wei.Analysis and application of perceptual weighting for AVS-M audio coder.《2007 3rd International Conference on Wireless Communications, Networking, and Mobile Computing- WiCOM 07》.2007, *
Zhang Wei.Analysis and application of perceptual weighting for AVS-M audio coder.《2007 3rd International Conference on Wireless Communications, Networking, and Mobile Computing - WiCOM 07》.2007,
Zhang Yong
蒋树强
郑轶佳
高文
黄庆明
黄庆明;郑轶佳;蒋树强;高文;.基于用户关注空间与注意力分析的视频精彩摘要与排序.《计算机学报》.2008, *

Also Published As

Publication number Publication date
CN101604526A (en) 2009-12-16

Similar Documents

Publication Publication Date Title
Lo et al. Mosnet: Deep learning based objective assessment for voice conversion
Le et al. Discretized Continuous Speech Emotion Recognition with Multi-Task Deep Recurrent Neural Network.
CN105529028B (en) Speech analysis method and apparatus
Armitage et al. A comparison of supervised learning techniques in the classification of bat echolocation calls
Morrison Measuring the validity and reliability of forensic likelihood-ratio systems
US8655655B2 (en) Sound event detecting module for a sound event recognition system and method thereof
Li et al. An automated assessment framework for atypical prosody and stereotyped idiosyncratic phrases related to autism spectrum disorder
Skowronski et al. Acoustic detection and classification of microchiroptera using machine learning: lessons learned from automatic speech recognition
CN102915728B (en) Sound segmentation device and method and speaker recognition system
US20140343935A1 (en) Apparatus and method for performing asynchronous speech recognition using multiple microphones
CN109712641A (en) A kind of processing method of audio classification and segmentation based on support vector machines
CN103440873A (en) Music recommendation method based on similarities
US20170004848A1 (en) Method for determining alcohol consumption, and recording medium and terminal for carrying out same
US10002623B2 (en) Speech-processing apparatus and speech-processing method
Gingras et al. A three-parameter model for classifying anurans into four genera based on advertisement calls
CN103489445A (en) Method and device for recognizing human voices in audio
CN112259101A (en) Voice keyword recognition method and device, computer equipment and storage medium
Iqbal et al. General-purpose audio tagging from noisy labels using convolutional neural networks.
Strömbergsson et al. Acoustic and perceptual evaluation of category goodness of/t/and/k/in typical and misarticulated children's speech
Lowell The acoustic assessment of voice in continuous speech
Wei et al. RMVPE: A robust model for vocal pitch estimation in polyphonic music
CN101604526B (en) Weight-based system and method for calculating audio frequency attention
CN107507627B (en) Voice data heat analysis method and system
CN116594057B (en) Earthquake early warning method and device based on deep learning and edge calculation
CN110739006B (en) Audio processing method and device, storage medium and electronic equipment

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20111116

CF01 Termination of patent right due to non-payment of annual fee