CN103971702A - Sound monitoring method, device and system - Google Patents

Sound monitoring method, device and system Download PDF

Info

Publication number
CN103971702A
CN103971702A CN201310332073.6A CN201310332073A CN103971702A CN 103971702 A CN103971702 A CN 103971702A CN 201310332073 A CN201310332073 A CN 201310332073A CN 103971702 A CN103971702 A CN 103971702A
Authority
CN
China
Prior art keywords
sound
training
detected
event model
voice signal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201310332073.6A
Other languages
Chinese (zh)
Inventor
何勇军
孙广路
谢怡宁
刘嘉辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Harbin University of Science and Technology
Original Assignee
Harbin University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Harbin University of Science and Technology filed Critical Harbin University of Science and Technology
Priority to CN201310332073.6A priority Critical patent/CN103971702A/en
Publication of CN103971702A publication Critical patent/CN103971702A/en
Pending legal-status Critical Current

Links

Landscapes

  • Measurement Of Mechanical Vibrations Or Ultrasonic Waves (AREA)

Abstract

The invention provides a sound monitoring method, device and system and relates to the technical field of sound signal processing and sound mode recognition. The method includes a sound training stage and a sound detecting stage. The sound training stage includes the steps of S1, acquiring sound training signals, and extracting sound training features; S2, training sound event models according to the sound training features. The sound detecting stage includes the steps of S3, extracting the features of to-be-detected sounds; S4, judging whether at least one of the sound event models is matched with the features of the to-be-detected sounds or not, if so, judging that a violent event exists, and if not, judging that no violent event exists. The method has the advantages that the sound features of sound signals are extracted and compared with trained sound event models, so that whether violent events exist in an elevator can be known through analyzing, automatic monitoring of the violent events in the elevator is achieved, the monitoring results can be provided in real time, and detecting accuracy can be effectively guaranteed.

Description

Sound monitoring method, Apparatus and system
Technical field
The present invention relates to sound signal processing and mode identification technology, be specifically related to a kind of sound monitoring method, Apparatus and system.
Background technology
Along with the high speed development of modern city, the use of elevator is more and more general and become the requisite vertical transportation instrument of skyscraper, closely related with resident's routine work and life.According to the statistics made by the departments concerned, up to now, the annual requirement of China's elevator has reached 1/3rd of the whole world.Meanwhile, due to elevator relative closure, become offender and implemented malfeasant splendid place, this has brought numerous potential safety hazards for daily life.Increasing offender implements to plunder, kill a person or sexual harassment in elevator, and elevator user's the security of the lives and property in serious threat.Document shows, elevator incident of violence in recent years presents the trend of increasing rapidly, only 2012 1 year, has the elevator crime dramas of placing on record just up to more than 6.2 ten thousand.Therefore, event in elevator is carried out to effectively monitoring and will the discovery of elevator incident of violence, prevention and detection etc. be had to important realistic meaning undoubtedly.
Extensively adopt at present the mode of camera video monitoring to realize effective monitoring to the incident of violence in elevator.
Although obtained certain effect, still existed following problem: the intelligent degree of monitoring is low, depend on Control Room staff's observation or the video of leafing through is found incident of violence.Obviously, this monitor mode will expend a large amount of manpower and materials and people sees that video image surpasses 20 minutes its notices and will obviously decline, and accuracy rate is also had a greatly reduced quality.
Summary of the invention
(1) technical matters solving
For the deficiencies in the prior art, the invention provides a kind of sound monitoring method, Apparatus and system, can automatically realize the monitoring of incident of violence in elevator.
(2) technical scheme
For realizing above object, the present invention is achieved by the following technical programs:
A sound monitoring method, comprises the training sound stage and detects the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
S2, according to described training sound characteristic, training sound event model;
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
Preferably, in step S1, comprise step:
S11, the described voice signal obtaining is carried out to pre-service;
S12, the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
S13, based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
S14, the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
Preferably, the pre-service in step S11 comprises minute frame operation and windowing operation;
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
The expression formula X that asks power spectrum described in step S12 a(k) be:
X a ( k ) = | | Σ n = 0 N - 1 x ( n ) e - j 2 kπ / N | | 2 , 0 ≤ k ≤ N
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
Preferably, in step S2, by gauss hybrid models, train sound incident of violence model, the probability density function of described M rank gauss hybrid models is as follows:
P ( o | λ ) = Σ i = 1 M c i P ( o | i , λ )
Wherein, P ( o | i , λ ) = N ( o , μ i , Σ i ) = 1 ( 2 π ) K / 2 | Σ i | 1 / 2 exp { - ( o - μ i ) T Σ i - 1 ( o - μ i ) 2 }
In formula, λ={ c i, μ i, Σ i; (i=1...M) }, μ ifor mean value vector, Σ ifor covariance matrix, i=1,2 ..M.Matrix Σ iadopt diagonal matrix here:
c i = 1 T Σ i = 1 T P ( q t = i | o t , λ )
μ i = Σ t = 1 T P ( q t = i | o t , λ ) o t Σ t = 1 T P ( q t = i | o t , λ )
Preferably, step S4 comprises following steps:
S31, supposition sound event model have N, and each sound event model, by a gauss hybrid models modeling, is respectively λ 1, λ 2..., λ n, in the judgement stage, the described sound characteristic to be detected of the observation of input integrates as O={o 1, o 2..., o t, the frame number that T is sound import;
S32, calculate the posterior probability that described sound to be detected is n sound event model, 1≤n≤N;
S33, according to described posterior probability, obtain anticipation result;
S34, according to described anticipation result, obtain final court verdict.
Preferably,
Calculating posterior probability expression formula in step S32 is:
p ( λ n | O ) = p ( O | λ n ) p ( λ n ) p ( O )
= p ( O | λ n ) p ( λ n ) Σ m = 1 N p ( O | λ m ) p ( λ m )
P ( λ n ) = 1 N , n = 1,2 , N .
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (O| λ n) be the conditional probability that n sound event model produces described sound characteristic collection O to be detected.
Preferably, the calculating anticipation result expression in step S33 is:
n * = arg max 1 ≤ n ≤ N ln P ( λ n | O ) = arg max 1 ≤ n ≤ N Σ t = 1 T ln P ( λ n | o t )
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (λ n| o t) be o tresult from λ nprobability;
Preferably, the calculating court verdict expression formula in step S34 is:
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; for o tresult from probability; Threshold refuses to know thresholding for default.
The present invention is a kind of sound monitoring device also, comprises with lower module:
Training sound stage module, obtains training voice signal, extracts the training sound characteristic of described training voice signal; According to described training sound characteristic, training sound event model;
Detect sound stage module, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected; Judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
The present invention also provides a kind of sound monitoring system, it is characterized in that, comprises microphone, and multiplexed signal sampling device also comprises sound monitoring device;
Described microphone is installed in elevator, and collected sound signal sends multiplexed signal sampling device to;
Described multiplexed signal sampling device, receives the voice signal that microphone sends, and sends sound monitoring device to;
Described sound monitoring device is processed voice signal.
(3) beneficial effect
The present invention, by a kind of sound monitoring method, Apparatus and system are provided, trains the training sound characteristic of voice signal by extraction, training sound event model; By extracting the sound characteristic to be detected of voice signal to be detected, extracted sound characteristic to be detected and training sound event model are compared, analysis draws in elevator, whether there is incident of violence, realized the automatic monitoring of incident of violence in elevator, provide in real time monitored results, can effectively guarantee the accuracy rate detecting, for next step processing of monitor staff provides foundation.
Equipment of the present invention is compared with the needed industrial camera of video monitoring, and microphone and relevant collecting device thereof have advantage with low cost, are convenient to promote the use of.
Microphone of the present invention is compared and the needed industrial camera of video monitoring, and volume is little, is convenient to be arranged in hiding corner, avoids being subject to offender's destruction, makes watch-dog safer.
Microphone of the present invention is compared and the needed industrial camera of video monitoring, and collection signal is not subject to illumination, block and the impact of the factor such as camouflage, makes monitor mode more stable.
Accompanying drawing explanation
In order to be illustrated more clearly in the embodiment of the present invention or technical scheme of the prior art, to the accompanying drawing of required use in embodiment or description of the Prior Art be briefly described below, apparently, accompanying drawing in the following describes is only some embodiments of the present invention, for those of ordinary skills, do not paying under the prerequisite of creative work, can also obtain according to these accompanying drawings other accompanying drawing.
Fig. 1 is the schematic flow sheet of a kind of sound monitoring method of the preferred embodiment of the present invention;
Fig. 2 is the schematic flow sheet of a kind of sound monitoring method of the preferred embodiment of the present invention;
Fig. 3 is the structural representation of a kind of sound monitoring device of the preferred embodiment of the present invention;
Fig. 4 is the configuration diagram of a kind of sound monitoring system of the preferred embodiment of the present invention.
Embodiment
Under regard to a kind of sound monitoring method, Apparatus and system proposed by the invention, describe in detail in conjunction with the accompanying drawings and embodiments.
Embodiment 1:
As shown in Figure 1, a kind of sound monitoring method, comprises the training sound stage and detects the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
S2, according to described training sound characteristic training sound event model;
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
The embodiment of the present invention, by a kind of sound monitoring method is provided, is trained the training sound characteristic of voice signal by extraction, training sound event model; By extracting the sound characteristic to be detected of voice signal to be detected, extracted sound characteristic to be detected and training sound event model are compared, analysis draws in elevator, whether there is incident of violence, realized the automatic monitoring of incident of violence in elevator, provide in real time monitored results, can effectively guarantee the accuracy rate detecting, for next step processing of monitor staff provides foundation.
The embodiment of the present invention is proceeded to detailed elaboration below:
A sound monitoring method, comprises the training sound stage and detects the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
Preferably, in step S1, comprise step:
S11, the training voice signal obtaining is carried out to pre-service;
S12, the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
S13, based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
S14, the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
Preferably, the pre-service in step S11 comprises minute frame operation and windowing operation;
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
Preferably, the expression formula X that asks power spectrum described in step S12 a(k) be:
X a ( k ) = | | Σ n = 0 N - 1 x ( n ) e - j 2 kπ / N | | 2 , 0 ≤ k ≤ N
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
S2, according to described training sound characteristic training sound event model;
The embodiment of the present invention is trained voice signal for each and is set up a GMM.The probability density function of M rank GMM is as follows:
P ( o | λ ) = Σ i = 1 M P ( o | i , λ ) = Σ i = 1 M c i P ( o | i , λ )
Wherein, λ is the parameter set of GMM model; O is the acoustic feature vector of K dimension; I is hidden state number, the sequence number of gaussian component namely, and M rank GMM just has M hidden state; c ibe the mixed weight-value of i component, its value corresponds to the prior probability of hidden state i, therefore has:
Σ i = 1 M c i = 1
P (o|i, λ) is Gaussian mixture components, the observation probability density function of corresponding hidden state i,
Wherein, in step S2, by gauss hybrid models, train sound incident of violence model, the probability density function of described M rank gauss hybrid models is as follows:
P ( o | λ ) = Σ i = 1 M c i P ( o | i , λ )
Wherein, P ( o | i , λ ) = N ( o , μ i , Σ i ) = 1 ( 2 π ) K / 2 | Σ i | 1 / 2 exp { - ( o - μ i ) T Σ i - 1 ( o - μ i ) 2 }
In formula, λ={ c i, μ i, Σ i; (i=1...M) }, μ ifor mean value vector, Σ ifor covariance matrix, i=1,2 ..M.Matrix Σ iadopt diagonal matrix here:
c i = 1 T Σ i = 1 T P ( q t = i | o t , λ )
μ i = Σ t = 1 T P ( q t = i | o t , λ ) o t Σ t = 1 T P ( q t = i | o t , λ )
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
Preferably, in step S3, comprise step:
S11 ', the voice signal to be detected obtaining is carried out to pre-service;
Preferably, the pre-service in step S11 ' comprises minute frame operation and windowing operation;
Wherein, divide the object of frame to be time signal to be divided into overlapping voice snippet, i.e. frame mutually.Every frame length is generally 30ms left and right, and frame moves as 10ms.
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
S12 ', the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
Preferably, the expression formula X that asks power spectrum described in step S12 ' a(k) be:
X a ( k ) = | | Σ n = 0 N - 1 x ( n ) e - j 2 kπ / N | | 2 , 0 ≤ k ≤ N
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
S13 ', based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
A bank of filters (number of wave filter is close with the number of critical band) that has M wave filter of embodiment of the present invention definition, the wave filter of employing is triangular filter, centre frequency is f (m), m=0,2 ..., M-1, the embodiment of the present invention is got M=28.In bank of filters, the span of each triangular filter equates in Mel scale, and the frequency response of triangular filter is defined as:
H m ( k ) = 0 k < f ( m - 1 ) ork > f ( m + 1 ) 2 ( k - f ( m - 1 ) ) ( f ( m + 1 ) - f ( m - 1 ) ) ( f ( m ) - f ( m - 1 ) ) f ( m - 1 ) < k < f ( m ) 2 ( f ( m + 1 ) - k ) ( f ( m + 1 ) - f ( m - 1 ) ) ( f ( m + 1 ) - f ( m ) ) f ( m ) &le; k &le; f ( m + 1 )
Next power spectrum is added to Mel bank of filters:
S ( m ) = ln ( &Sigma; k = 0 N - 1 | X a ( k ) | 2 H m ( k ) ) , 0 &le; m &le; M
Then do discrete cosine transform (DCT) and obtain Mel cepstrum coefficient:
c ( n ) = &Sigma; m = 0 M - 1 S ( m ) cos ( n&pi; ( m - 0.5 ) / M ) , 0 &le; n &le; M
S14 ', the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
If t and t+1 cepstrum vector is constantly c tand c t+1,
The computing method of first order difference are:
Δc t=c t+1-c t
Second order difference is:
ΔΔc t=Δc t+1-Δc t
Spliced phonetic feature is:
[c tΔc tΔΔc t]
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
Preferably, step S4 comprises following steps:
S31, supposition sound event model have N, and each sound event model, by a gauss hybrid models modeling, is respectively λ 1, λ 2..., λ n, in the judgement stage, the described sound characteristic to be detected of the observation of input integrates as O={o 1, o 2..., o t, the frame number that T is sound import;
S32, calculate the posterior probability that described sound to be detected is n sound event model, 1≤n≤N;
S33, according to described posterior probability, obtain anticipation result;
S34, according to described anticipation result, obtain final court verdict.
Preferably, the calculating posterior probability expression formula in step S32 is:
p ( &lambda; n | O ) = p ( O | &lambda; n ) p ( &lambda; n ) p ( O )
= p ( O | &lambda; n ) p ( &lambda; n ) &Sigma; m = 1 N p ( O | &lambda; m ) p ( &lambda; m )
P ( &lambda; n ) = 1 N , n = 1,2 , N .
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (O| λ n) be the conditional probability that n sound event model produces described sound characteristic collection O to be detected.
Preferably, the calculating anticipation result expression in step S33 is:
n * = arg max 1 &le; n &le; N ln P ( &lambda; n | O ) = arg max 1 &le; n &le; N &Sigma; t = 1 T ln P ( &lambda; n | o t )
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (λ n| o t) be o tresult from λ nprobability;
Preferably, the calculating court verdict expression formula in step S34 is:
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; for o tresult from probability; Threshold refuses to know thresholding for default.
Embodiment 2:
As shown in Figure 3, a kind of sound monitoring device, is characterized in that, comprises with lower module:
Training sound stage module, obtains training voice signal, extracts the training sound characteristic of described training voice signal; According to described training sound characteristic, training sound event model;
Detect sound stage module, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected; Judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
Embodiment 3:
As shown in Figure 4, a kind of sound monitoring system, is characterized in that, comprises microphone, and multiplexed signal sampling device also comprises sound monitoring device as described in example 2 above;
Described microphone is installed in elevator, and collected sound signal sends multiplexed signal sampling device to;
Described multiplexed signal sampling device, receives the voice signal that microphone sends, and sends sound monitoring device to;
Described sound monitoring device is processed voice signal.
To sum up, the embodiment of the present invention, by a kind of sound monitoring method, Apparatus and system are provided, is trained the training sound characteristic of voice signal by extraction, training sound event model; By extracting the sound characteristic to be detected of voice signal to be detected, extracted sound characteristic to be detected and training sound event model are compared, analysis draws in elevator, whether there is incident of violence, realized the automatic monitoring of incident of violence in elevator, provide in real time monitored results, can effectively guarantee the accuracy rate detecting, for next step processing of monitor staff provides foundation.
The equipment that the embodiment of the present invention adopts is compared with the needed industrial camera of video monitoring, and microphone and relevant collecting device thereof have advantage with low cost, are convenient to promote the use of.
The microphone that the embodiment of the present invention adopts is compared and the needed industrial camera of video monitoring, and volume is little, is convenient to be arranged in hiding corner, avoids being subject to offender's destruction, makes watch-dog safer.
The microphone that the embodiment of the present invention adopts is compared and the needed industrial camera of video monitoring, and collection signal is not subject to illumination, block and the impact of the factor such as camouflage, makes monitor mode more stable.
It should be noted that, in this article, term " comprises ", " comprising " or its any other variant are intended to contain comprising of nonexcludability, thereby the process, method, article or the equipment that make to comprise a series of key elements not only comprise those key elements, but also comprise other key elements of clearly not listing, or be also included as the intrinsic key element of this process, method, article or equipment.The in the situation that of more restrictions not, the key element being limited by statement " comprising ... ", and be not precluded within process, method, article or the equipment that comprises described key element and also have other identical element.
Above embodiment only, in order to technical scheme of the present invention to be described, is not intended to limit; Although the present invention is had been described in detail with reference to previous embodiment, those of ordinary skill in the art is to be understood that: its technical scheme that still can record aforementioned each embodiment is modified, or part technical characterictic is wherein equal to replacement; And these modifications or replacement do not make the essence of appropriate technical solution depart from the spirit and scope of various embodiments of the present invention technical scheme.

Claims (10)

1. a sound monitoring method, is characterized in that, comprise the training sound stage and detect the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
S2, according to described training sound characteristic training sound event model;
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
2. sound monitoring method as claimed in claim 1, is characterized in that, comprises step in step S1 or step S3:
S11, the voice signal obtaining is carried out to pre-service;
S12, the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
S13, based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
S14, the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
3. sound monitoring method as claimed in claim 2, is characterized in that,
Pre-service in step S11 comprises minute frame operation and windowing operation;
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
The expression formula X that asks power spectrum described in step S12 a(k) be:
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
4. sound monitoring method as claimed in claim 1, is characterized in that, in step S2, by gauss hybrid models, trains sound incident of violence model, and the probability density function of described M rank gauss hybrid models is as follows:
Wherein,
In formula, λ={ c i, μ i, Σ i; (i=1...M) }, μ ifor mean value vector, Σ ifor covariance matrix, i=1,2 ..M.Matrix Σ iadopt diagonal matrix here:
5. sound monitoring method as claimed in claim 1, is characterized in that, step S4 comprises following steps:
S31, supposition sound event model have N, and each sound event model, by a gauss hybrid models modeling, is respectively λ 1, λ 2..., λ n, in the judgement stage, the described sound characteristic to be detected of the observation of input integrates as O={o 1, o 2..., o t, the frame number that T is sound import;
S32, calculate the posterior probability that described sound to be detected is n sound event model, 1≤n≤N;
S33, according to described posterior probability, obtain anticipation result;
S34, according to described anticipation result, obtain final court verdict.
6. sound monitoring method as claimed in claim 5, is characterized in that,
Calculating posterior probability expression formula in step S32 is:
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (O| λ n) be the conditional probability that n sound event model produces described sound characteristic collection O to be detected.
7. sound monitoring method as claimed in claim 5, is characterized in that,
Calculating anticipation result expression in step S33 is:
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (λ n| o t) be o tresult from λ nprobability.
8. sound monitoring method as claimed in claim 5, is characterized in that,
Calculating court verdict expression formula in step S34 is:
In formula, p (λ n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; for o tresult from probability; Threshold refuses to know thresholding for default.
9. a sound monitoring device, is characterized in that, comprises with lower module:
Training sound stage module, obtains training voice signal, extracts the training sound characteristic of described training voice signal; According to described training sound characteristic, training sound event model;
Detect sound stage module, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected; Judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
10. a sound monitoring system, is characterized in that, comprises microphone, and multiplexed signal sampling device also comprises sound monitoring device as claimed in claim 9;
Described microphone is installed in elevator, and collected sound signal sends multiplexed signal sampling device to;
Described multiplexed signal sampling device, receives the voice signal that microphone sends, and sends sound monitoring device to;
Described sound monitoring device is processed voice signal.
CN201310332073.6A 2013-08-01 2013-08-01 Sound monitoring method, device and system Pending CN103971702A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310332073.6A CN103971702A (en) 2013-08-01 2013-08-01 Sound monitoring method, device and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310332073.6A CN103971702A (en) 2013-08-01 2013-08-01 Sound monitoring method, device and system

Publications (1)

Publication Number Publication Date
CN103971702A true CN103971702A (en) 2014-08-06

Family

ID=51241116

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310332073.6A Pending CN103971702A (en) 2013-08-01 2013-08-01 Sound monitoring method, device and system

Country Status (1)

Country Link
CN (1) CN103971702A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105679313A (en) * 2016-04-15 2016-06-15 福建新恒通智能科技有限公司 Audio recognition alarm system and method
CN107527617A (en) * 2017-09-30 2017-12-29 上海应用技术大学 Monitoring method, apparatus and system based on voice recognition
CN107910019A (en) * 2017-11-30 2018-04-13 中国科学院微电子研究所 A kind of human acoustical signal's processing and analysis method
CN110223715A (en) * 2019-05-07 2019-09-10 华南理工大学 It is a kind of based on sound event detection old solitary people man in activity estimation method
CN110800053A (en) * 2017-06-13 2020-02-14 米纳特有限公司 Method and apparatus for obtaining event indications based on audio data
CN111326172A (en) * 2018-12-17 2020-06-23 北京嘀嘀无限科技发展有限公司 Conflict detection method and device, electronic equipment and readable storage medium
WO2020140552A1 (en) * 2018-12-31 2020-07-09 瑞声声学科技(深圳)有限公司 Haptic feedback method
CN111599379A (en) * 2020-05-09 2020-08-28 北京南师信息技术有限公司 Conflict early warning method, device, equipment, readable storage medium and triage system
CN113421544A (en) * 2021-06-30 2021-09-21 平安科技(深圳)有限公司 Singing voice synthesis method and device, computer equipment and storage medium
CN113670434A (en) * 2021-06-21 2021-11-19 深圳供电局有限公司 Transformer substation equipment sound abnormality identification method and device and computer equipment

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101477798A (en) * 2009-02-17 2009-07-08 北京邮电大学 Method for analyzing and extracting audio data of set scene
CN101587710A (en) * 2009-07-02 2009-11-25 北京理工大学 A kind of many code books coding parameter quantification method based on the audio emergent event classification
CN102509545A (en) * 2011-09-21 2012-06-20 哈尔滨工业大学 Real time acoustics event detecting system and method
CN102799899A (en) * 2012-06-29 2012-11-28 北京理工大学 Special audio event layered and generalized identification method based on SVM (Support Vector Machine) and GMM (Gaussian Mixture Model)
CN103177722A (en) * 2013-03-08 2013-06-26 北京理工大学 Tone-similarity-based song retrieval method
CN103226948A (en) * 2013-04-22 2013-07-31 山东师范大学 Audio scene recognition method based on acoustic events

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101477798A (en) * 2009-02-17 2009-07-08 北京邮电大学 Method for analyzing and extracting audio data of set scene
CN101587710A (en) * 2009-07-02 2009-11-25 北京理工大学 A kind of many code books coding parameter quantification method based on the audio emergent event classification
CN102509545A (en) * 2011-09-21 2012-06-20 哈尔滨工业大学 Real time acoustics event detecting system and method
CN102799899A (en) * 2012-06-29 2012-11-28 北京理工大学 Special audio event layered and generalized identification method based on SVM (Support Vector Machine) and GMM (Gaussian Mixture Model)
CN103177722A (en) * 2013-03-08 2013-06-26 北京理工大学 Tone-similarity-based song retrieval method
CN103226948A (en) * 2013-04-22 2013-07-31 山东师范大学 Audio scene recognition method based on acoustic events

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
蒋刚 等: "《工业机器人》", 31 January 2011 *
韩纪庆 等: "《音频信息检索理论与技术》", 31 March 2011 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105679313A (en) * 2016-04-15 2016-06-15 福建新恒通智能科技有限公司 Audio recognition alarm system and method
CN110800053A (en) * 2017-06-13 2020-02-14 米纳特有限公司 Method and apparatus for obtaining event indications based on audio data
CN107527617A (en) * 2017-09-30 2017-12-29 上海应用技术大学 Monitoring method, apparatus and system based on voice recognition
CN107910019A (en) * 2017-11-30 2018-04-13 中国科学院微电子研究所 A kind of human acoustical signal's processing and analysis method
CN111326172A (en) * 2018-12-17 2020-06-23 北京嘀嘀无限科技发展有限公司 Conflict detection method and device, electronic equipment and readable storage medium
WO2020140552A1 (en) * 2018-12-31 2020-07-09 瑞声声学科技(深圳)有限公司 Haptic feedback method
CN110223715A (en) * 2019-05-07 2019-09-10 华南理工大学 It is a kind of based on sound event detection old solitary people man in activity estimation method
CN110223715B (en) * 2019-05-07 2021-05-25 华南理工大学 Home activity estimation method for solitary old people based on sound event detection
CN111599379A (en) * 2020-05-09 2020-08-28 北京南师信息技术有限公司 Conflict early warning method, device, equipment, readable storage medium and triage system
CN111599379B (en) * 2020-05-09 2023-09-29 北京南师信息技术有限公司 Conflict early warning method, device, equipment, readable storage medium and triage system
CN113670434A (en) * 2021-06-21 2021-11-19 深圳供电局有限公司 Transformer substation equipment sound abnormality identification method and device and computer equipment
CN113421544A (en) * 2021-06-30 2021-09-21 平安科技(深圳)有限公司 Singing voice synthesis method and device, computer equipment and storage medium
CN113421544B (en) * 2021-06-30 2024-05-10 平安科技(深圳)有限公司 Singing voice synthesizing method, singing voice synthesizing device, computer equipment and storage medium

Similar Documents

Publication Publication Date Title
CN103971702A (en) Sound monitoring method, device and system
CN102664006B (en) Abnormal voice detecting method based on time-domain and frequency-domain analysis
CN106504754B (en) A kind of real-time method for generating captions according to audio output
CN104732978B (en) The relevant method for distinguishing speek person of text based on combined depth study
CN109616140B (en) Abnormal sound analysis system
CN103971700A (en) Voice monitoring method and device
CN107527617A (en) Monitoring method, apparatus and system based on voice recognition
CN101494049A (en) Method for extracting audio characteristic parameter of audio monitoring system
CN105841797A (en) Window motor abnormal noise detection method and apparatus based on MFCC and SVM
CN103117061A (en) Method and device for identifying animals based on voice
Schröder et al. Classification of human cough signals using spectro-temporal Gabor filterbank features
CN104732972B (en) A kind of HMM Application on Voiceprint Recognition based on classified statistics is registered method and system
CN108520753A (en) Voice lie detection method based on the two-way length of convolution memory network in short-term
CN105608823A (en) Optical fiber security and protection method and system based on principal component analysis
CN110522462A (en) The multi-modal intelligent trial system of one kind and method
CN115102789A (en) Anti-communication network fraud studying, judging, early-warning and intercepting comprehensive platform
CN103021421A (en) Multilevel screening detecting recognizing method for shots
CN115910097A (en) Audible signal identification method and system for latent fault of high-voltage circuit breaker
CN110931024B (en) Audio-based prediction method and system for natural mating result of captive pandas
CN104091104B (en) Multi-format audio perceives the characteristics extraction of Hash certification and authentication method
CN109389994A (en) Identification of sound source method and device for intelligent transportation system
CN115854269A (en) Leakage hole jet flow noise identification method and device, electronic equipment and storage medium
CN108470564A (en) According to the artificial intelligence approach of audio identification personality characteristics
CN111524523A (en) Instrument and equipment state detection system and method based on voiceprint recognition technology
CN106297805A (en) A kind of method for distinguishing speek person based on respiratory characteristic

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20140806