CN103971702A - Sound monitoring method, device and system - Google Patents
Sound monitoring method, device and system Download PDFInfo
- Publication number
- CN103971702A CN103971702A CN201310332073.6A CN201310332073A CN103971702A CN 103971702 A CN103971702 A CN 103971702A CN 201310332073 A CN201310332073 A CN 201310332073A CN 103971702 A CN103971702 A CN 103971702A
- Authority
- CN
- China
- Prior art keywords
- sound
- training
- detected
- event model
- voice signal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Landscapes
- Measurement Of Mechanical Vibrations Or Ultrasonic Waves (AREA)
Abstract
The invention provides a sound monitoring method, device and system and relates to the technical field of sound signal processing and sound mode recognition. The method includes a sound training stage and a sound detecting stage. The sound training stage includes the steps of S1, acquiring sound training signals, and extracting sound training features; S2, training sound event models according to the sound training features. The sound detecting stage includes the steps of S3, extracting the features of to-be-detected sounds; S4, judging whether at least one of the sound event models is matched with the features of the to-be-detected sounds or not, if so, judging that a violent event exists, and if not, judging that no violent event exists. The method has the advantages that the sound features of sound signals are extracted and compared with trained sound event models, so that whether violent events exist in an elevator can be known through analyzing, automatic monitoring of the violent events in the elevator is achieved, the monitoring results can be provided in real time, and detecting accuracy can be effectively guaranteed.
Description
Technical field
The present invention relates to sound signal processing and mode identification technology, be specifically related to a kind of sound monitoring method, Apparatus and system.
Background technology
Along with the high speed development of modern city, the use of elevator is more and more general and become the requisite vertical transportation instrument of skyscraper, closely related with resident's routine work and life.According to the statistics made by the departments concerned, up to now, the annual requirement of China's elevator has reached 1/3rd of the whole world.Meanwhile, due to elevator relative closure, become offender and implemented malfeasant splendid place, this has brought numerous potential safety hazards for daily life.Increasing offender implements to plunder, kill a person or sexual harassment in elevator, and elevator user's the security of the lives and property in serious threat.Document shows, elevator incident of violence in recent years presents the trend of increasing rapidly, only 2012 1 year, has the elevator crime dramas of placing on record just up to more than 6.2 ten thousand.Therefore, event in elevator is carried out to effectively monitoring and will the discovery of elevator incident of violence, prevention and detection etc. be had to important realistic meaning undoubtedly.
Extensively adopt at present the mode of camera video monitoring to realize effective monitoring to the incident of violence in elevator.
Although obtained certain effect, still existed following problem: the intelligent degree of monitoring is low, depend on Control Room staff's observation or the video of leafing through is found incident of violence.Obviously, this monitor mode will expend a large amount of manpower and materials and people sees that video image surpasses 20 minutes its notices and will obviously decline, and accuracy rate is also had a greatly reduced quality.
Summary of the invention
(1) technical matters solving
For the deficiencies in the prior art, the invention provides a kind of sound monitoring method, Apparatus and system, can automatically realize the monitoring of incident of violence in elevator.
(2) technical scheme
For realizing above object, the present invention is achieved by the following technical programs:
A sound monitoring method, comprises the training sound stage and detects the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
S2, according to described training sound characteristic, training sound event model;
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
Preferably, in step S1, comprise step:
S11, the described voice signal obtaining is carried out to pre-service;
S12, the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
S13, based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
S14, the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
Preferably, the pre-service in step S11 comprises minute frame operation and windowing operation;
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
The expression formula X that asks power spectrum described in step S12
a(k) be:
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
Preferably, in step S2, by gauss hybrid models, train sound incident of violence model, the probability density function of described M rank gauss hybrid models is as follows:
Wherein,
In formula, λ={ c
i, μ
i, Σ
i; (i=1...M) }, μ
ifor mean value vector, Σ
ifor covariance matrix, i=1,2 ..M.Matrix Σ
iadopt diagonal matrix here:
Preferably, step S4 comprises following steps:
S31, supposition sound event model have N, and each sound event model, by a gauss hybrid models modeling, is respectively λ
1, λ
2..., λ
n, in the judgement stage, the described sound characteristic to be detected of the observation of input integrates as O={o
1, o
2..., o
t, the frame number that T is sound import;
S32, calculate the posterior probability that described sound to be detected is n sound event model, 1≤n≤N;
S33, according to described posterior probability, obtain anticipation result;
S34, according to described anticipation result, obtain final court verdict.
Preferably,
Calculating posterior probability expression formula in step S32 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (O| λ
n) be the conditional probability that n sound event model produces described sound characteristic collection O to be detected.
Preferably, the calculating anticipation result expression in step S33 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (λ
n| o
t) be o
tresult from λ
nprobability;
Preferably, the calculating court verdict expression formula in step S34 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions;
for o
tresult from
probability; Threshold refuses to know thresholding for default.
The present invention is a kind of sound monitoring device also, comprises with lower module:
Training sound stage module, obtains training voice signal, extracts the training sound characteristic of described training voice signal; According to described training sound characteristic, training sound event model;
Detect sound stage module, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected; Judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
The present invention also provides a kind of sound monitoring system, it is characterized in that, comprises microphone, and multiplexed signal sampling device also comprises sound monitoring device;
Described microphone is installed in elevator, and collected sound signal sends multiplexed signal sampling device to;
Described multiplexed signal sampling device, receives the voice signal that microphone sends, and sends sound monitoring device to;
Described sound monitoring device is processed voice signal.
(3) beneficial effect
The present invention, by a kind of sound monitoring method, Apparatus and system are provided, trains the training sound characteristic of voice signal by extraction, training sound event model; By extracting the sound characteristic to be detected of voice signal to be detected, extracted sound characteristic to be detected and training sound event model are compared, analysis draws in elevator, whether there is incident of violence, realized the automatic monitoring of incident of violence in elevator, provide in real time monitored results, can effectively guarantee the accuracy rate detecting, for next step processing of monitor staff provides foundation.
Equipment of the present invention is compared with the needed industrial camera of video monitoring, and microphone and relevant collecting device thereof have advantage with low cost, are convenient to promote the use of.
Microphone of the present invention is compared and the needed industrial camera of video monitoring, and volume is little, is convenient to be arranged in hiding corner, avoids being subject to offender's destruction, makes watch-dog safer.
Microphone of the present invention is compared and the needed industrial camera of video monitoring, and collection signal is not subject to illumination, block and the impact of the factor such as camouflage, makes monitor mode more stable.
Accompanying drawing explanation
In order to be illustrated more clearly in the embodiment of the present invention or technical scheme of the prior art, to the accompanying drawing of required use in embodiment or description of the Prior Art be briefly described below, apparently, accompanying drawing in the following describes is only some embodiments of the present invention, for those of ordinary skills, do not paying under the prerequisite of creative work, can also obtain according to these accompanying drawings other accompanying drawing.
Fig. 1 is the schematic flow sheet of a kind of sound monitoring method of the preferred embodiment of the present invention;
Fig. 2 is the schematic flow sheet of a kind of sound monitoring method of the preferred embodiment of the present invention;
Fig. 3 is the structural representation of a kind of sound monitoring device of the preferred embodiment of the present invention;
Fig. 4 is the configuration diagram of a kind of sound monitoring system of the preferred embodiment of the present invention.
Embodiment
Under regard to a kind of sound monitoring method, Apparatus and system proposed by the invention, describe in detail in conjunction with the accompanying drawings and embodiments.
Embodiment 1:
As shown in Figure 1, a kind of sound monitoring method, comprises the training sound stage and detects the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
S2, according to described training sound characteristic training sound event model;
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
The embodiment of the present invention, by a kind of sound monitoring method is provided, is trained the training sound characteristic of voice signal by extraction, training sound event model; By extracting the sound characteristic to be detected of voice signal to be detected, extracted sound characteristic to be detected and training sound event model are compared, analysis draws in elevator, whether there is incident of violence, realized the automatic monitoring of incident of violence in elevator, provide in real time monitored results, can effectively guarantee the accuracy rate detecting, for next step processing of monitor staff provides foundation.
The embodiment of the present invention is proceeded to detailed elaboration below:
A sound monitoring method, comprises the training sound stage and detects the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
Preferably, in step S1, comprise step:
S11, the training voice signal obtaining is carried out to pre-service;
S12, the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
S13, based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
S14, the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
Preferably, the pre-service in step S11 comprises minute frame operation and windowing operation;
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
Preferably, the expression formula X that asks power spectrum described in step S12
a(k) be:
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
S2, according to described training sound characteristic training sound event model;
The embodiment of the present invention is trained voice signal for each and is set up a GMM.The probability density function of M rank GMM is as follows:
Wherein, λ is the parameter set of GMM model; O is the acoustic feature vector of K dimension; I is hidden state number, the sequence number of gaussian component namely, and M rank GMM just has M hidden state; c
ibe the mixed weight-value of i component, its value corresponds to the prior probability of hidden state i, therefore has:
P (o|i, λ) is Gaussian mixture components, the observation probability density function of corresponding hidden state i,
Wherein, in step S2, by gauss hybrid models, train sound incident of violence model, the probability density function of described M rank gauss hybrid models is as follows:
Wherein,
In formula, λ={ c
i, μ
i, Σ
i; (i=1...M) }, μ
ifor mean value vector, Σ
ifor covariance matrix, i=1,2 ..M.Matrix Σ
iadopt diagonal matrix here:
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
Preferably, in step S3, comprise step:
S11 ', the voice signal to be detected obtaining is carried out to pre-service;
Preferably, the pre-service in step S11 ' comprises minute frame operation and windowing operation;
Wherein, divide the object of frame to be time signal to be divided into overlapping voice snippet, i.e. frame mutually.Every frame length is generally 30ms left and right, and frame moves as 10ms.
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
S12 ', the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
Preferably, the expression formula X that asks power spectrum described in step S12 '
a(k) be:
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
S13 ', based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
A bank of filters (number of wave filter is close with the number of critical band) that has M wave filter of embodiment of the present invention definition, the wave filter of employing is triangular filter, centre frequency is f (m), m=0,2 ..., M-1, the embodiment of the present invention is got M=28.In bank of filters, the span of each triangular filter equates in Mel scale, and the frequency response of triangular filter is defined as:
Next power spectrum is added to Mel bank of filters:
Then do discrete cosine transform (DCT) and obtain Mel cepstrum coefficient:
S14 ', the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
If t and t+1 cepstrum vector is constantly c
tand c
t+1,
The computing method of first order difference are:
Δc
t=c
t+1-c
t
Second order difference is:
ΔΔc
t=Δc
t+1-Δc
t
Spliced phonetic feature is:
[c
tΔc
tΔΔc
t]
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
Preferably, step S4 comprises following steps:
S31, supposition sound event model have N, and each sound event model, by a gauss hybrid models modeling, is respectively λ
1, λ
2..., λ
n, in the judgement stage, the described sound characteristic to be detected of the observation of input integrates as O={o
1, o
2..., o
t, the frame number that T is sound import;
S32, calculate the posterior probability that described sound to be detected is n sound event model, 1≤n≤N;
S33, according to described posterior probability, obtain anticipation result;
S34, according to described anticipation result, obtain final court verdict.
Preferably, the calculating posterior probability expression formula in step S32 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (O| λ
n) be the conditional probability that n sound event model produces described sound characteristic collection O to be detected.
Preferably, the calculating anticipation result expression in step S33 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (λ
n| o
t) be o
tresult from λ
nprobability;
Preferably, the calculating court verdict expression formula in step S34 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions;
for o
tresult from
probability; Threshold refuses to know thresholding for default.
Embodiment 2:
As shown in Figure 3, a kind of sound monitoring device, is characterized in that, comprises with lower module:
Training sound stage module, obtains training voice signal, extracts the training sound characteristic of described training voice signal; According to described training sound characteristic, training sound event model;
Detect sound stage module, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected; Judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
Embodiment 3:
As shown in Figure 4, a kind of sound monitoring system, is characterized in that, comprises microphone, and multiplexed signal sampling device also comprises sound monitoring device as described in example 2 above;
Described microphone is installed in elevator, and collected sound signal sends multiplexed signal sampling device to;
Described multiplexed signal sampling device, receives the voice signal that microphone sends, and sends sound monitoring device to;
Described sound monitoring device is processed voice signal.
To sum up, the embodiment of the present invention, by a kind of sound monitoring method, Apparatus and system are provided, is trained the training sound characteristic of voice signal by extraction, training sound event model; By extracting the sound characteristic to be detected of voice signal to be detected, extracted sound characteristic to be detected and training sound event model are compared, analysis draws in elevator, whether there is incident of violence, realized the automatic monitoring of incident of violence in elevator, provide in real time monitored results, can effectively guarantee the accuracy rate detecting, for next step processing of monitor staff provides foundation.
The equipment that the embodiment of the present invention adopts is compared with the needed industrial camera of video monitoring, and microphone and relevant collecting device thereof have advantage with low cost, are convenient to promote the use of.
The microphone that the embodiment of the present invention adopts is compared and the needed industrial camera of video monitoring, and volume is little, is convenient to be arranged in hiding corner, avoids being subject to offender's destruction, makes watch-dog safer.
The microphone that the embodiment of the present invention adopts is compared and the needed industrial camera of video monitoring, and collection signal is not subject to illumination, block and the impact of the factor such as camouflage, makes monitor mode more stable.
It should be noted that, in this article, term " comprises ", " comprising " or its any other variant are intended to contain comprising of nonexcludability, thereby the process, method, article or the equipment that make to comprise a series of key elements not only comprise those key elements, but also comprise other key elements of clearly not listing, or be also included as the intrinsic key element of this process, method, article or equipment.The in the situation that of more restrictions not, the key element being limited by statement " comprising ... ", and be not precluded within process, method, article or the equipment that comprises described key element and also have other identical element.
Above embodiment only, in order to technical scheme of the present invention to be described, is not intended to limit; Although the present invention is had been described in detail with reference to previous embodiment, those of ordinary skill in the art is to be understood that: its technical scheme that still can record aforementioned each embodiment is modified, or part technical characterictic is wherein equal to replacement; And these modifications or replacement do not make the essence of appropriate technical solution depart from the spirit and scope of various embodiments of the present invention technical scheme.
Claims (10)
1. a sound monitoring method, is characterized in that, comprise the training sound stage and detect the sound stage,
The described training sound stage comprises step:
S1, obtain training voice signal, extract the training sound characteristic of described training voice signal;
S2, according to described training sound characteristic training sound event model;
The described detection sound stage comprises step:
S3, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected;
S4, judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
2. sound monitoring method as claimed in claim 1, is characterized in that, comprises step in step S1 or step S3:
S11, the voice signal obtaining is carried out to pre-service;
S12, the voice signal after pretreated is made to discrete Fourier transform (DFT), try to achieve power spectrum;
S13, based on Mel bank of filters, try to achieve the Mel cepstrum coefficient of described power spectrum;
S14, the first order difference of calculating described Mel cepstrum coefficient and second order difference, by the coefficient of described first order difference and second order difference and the splicing of described Mel cepstrum coefficient, form sound characteristic.
3. sound monitoring method as claimed in claim 2, is characterized in that,
Pre-service in step S11 comprises minute frame operation and windowing operation;
Wherein, the window function that windowing operation adopts is Hamming window, and expression formula w (n) is:
In formula, n is time sequence number, and L is that window is long;
The expression formula X that asks power spectrum described in step S12
a(k) be:
X in formula (n) is the speech frame after windowing, and N represents counting of Fourier transform, and j represents imaginary unit.
4. sound monitoring method as claimed in claim 1, is characterized in that, in step S2, by gauss hybrid models, trains sound incident of violence model, and the probability density function of described M rank gauss hybrid models is as follows:
Wherein,
In formula, λ={ c
i, μ
i, Σ
i; (i=1...M) }, μ
ifor mean value vector, Σ
ifor covariance matrix, i=1,2 ..M.Matrix Σ
iadopt diagonal matrix here:
5. sound monitoring method as claimed in claim 1, is characterized in that, step S4 comprises following steps:
S31, supposition sound event model have N, and each sound event model, by a gauss hybrid models modeling, is respectively λ
1, λ
2..., λ
n, in the judgement stage, the described sound characteristic to be detected of the observation of input integrates as O={o
1, o
2..., o
t, the frame number that T is sound import;
S32, calculate the posterior probability that described sound to be detected is n sound event model, 1≤n≤N;
S33, according to described posterior probability, obtain anticipation result;
S34, according to described anticipation result, obtain final court verdict.
6. sound monitoring method as claimed in claim 5, is characterized in that,
Calculating posterior probability expression formula in step S32 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (O| λ
n) be the conditional probability that n sound event model produces described sound characteristic collection O to be detected.
7. sound monitoring method as claimed in claim 5, is characterized in that,
Calculating anticipation result expression in step S33 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions; P (λ
n| o
t) be o
tresult from λ
nprobability.
8. sound monitoring method as claimed in claim 5, is characterized in that,
Calculating court verdict expression formula in step S34 is:
In formula, p (λ
n) be the prior probability of n sound event model; P (O) is the probability of described sound characteristic collection O to be detected under all sound event Model Conditions;
for o
tresult from
probability; Threshold refuses to know thresholding for default.
9. a sound monitoring device, is characterized in that, comprises with lower module:
Training sound stage module, obtains training voice signal, extracts the training sound characteristic of described training voice signal; According to described training sound characteristic, training sound event model;
Detect sound stage module, obtain voice signal to be detected, extract the sound characteristic to be detected of described voice signal to be detected; Judge the sound event model that whether exists at least one to mate with described sound characteristic to be detected in described sound event model, as be yes, judge and have incident of violence; As be no, judge and do not have incident of violence.
10. a sound monitoring system, is characterized in that, comprises microphone, and multiplexed signal sampling device also comprises sound monitoring device as claimed in claim 9;
Described microphone is installed in elevator, and collected sound signal sends multiplexed signal sampling device to;
Described multiplexed signal sampling device, receives the voice signal that microphone sends, and sends sound monitoring device to;
Described sound monitoring device is processed voice signal.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310332073.6A CN103971702A (en) | 2013-08-01 | 2013-08-01 | Sound monitoring method, device and system |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201310332073.6A CN103971702A (en) | 2013-08-01 | 2013-08-01 | Sound monitoring method, device and system |
Publications (1)
Publication Number | Publication Date |
---|---|
CN103971702A true CN103971702A (en) | 2014-08-06 |
Family
ID=51241116
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201310332073.6A Pending CN103971702A (en) | 2013-08-01 | 2013-08-01 | Sound monitoring method, device and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN103971702A (en) |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105679313A (en) * | 2016-04-15 | 2016-06-15 | 福建新恒通智能科技有限公司 | Audio recognition alarm system and method |
CN107527617A (en) * | 2017-09-30 | 2017-12-29 | 上海应用技术大学 | Monitoring method, apparatus and system based on voice recognition |
CN107910019A (en) * | 2017-11-30 | 2018-04-13 | 中国科学院微电子研究所 | A kind of human acoustical signal's processing and analysis method |
CN110223715A (en) * | 2019-05-07 | 2019-09-10 | 华南理工大学 | It is a kind of based on sound event detection old solitary people man in activity estimation method |
CN110800053A (en) * | 2017-06-13 | 2020-02-14 | 米纳特有限公司 | Method and apparatus for obtaining event indications based on audio data |
CN111326172A (en) * | 2018-12-17 | 2020-06-23 | 北京嘀嘀无限科技发展有限公司 | Conflict detection method and device, electronic equipment and readable storage medium |
WO2020140552A1 (en) * | 2018-12-31 | 2020-07-09 | 瑞声声学科技(深圳)有限公司 | Haptic feedback method |
CN111599379A (en) * | 2020-05-09 | 2020-08-28 | 北京南师信息技术有限公司 | Conflict early warning method, device, equipment, readable storage medium and triage system |
CN113421544A (en) * | 2021-06-30 | 2021-09-21 | 平安科技(深圳)有限公司 | Singing voice synthesis method and device, computer equipment and storage medium |
CN113670434A (en) * | 2021-06-21 | 2021-11-19 | 深圳供电局有限公司 | Transformer substation equipment sound abnormality identification method and device and computer equipment |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101477798A (en) * | 2009-02-17 | 2009-07-08 | 北京邮电大学 | Method for analyzing and extracting audio data of set scene |
CN101587710A (en) * | 2009-07-02 | 2009-11-25 | 北京理工大学 | A kind of many code books coding parameter quantification method based on the audio emergent event classification |
CN102509545A (en) * | 2011-09-21 | 2012-06-20 | 哈尔滨工业大学 | Real time acoustics event detecting system and method |
CN102799899A (en) * | 2012-06-29 | 2012-11-28 | 北京理工大学 | Special audio event layered and generalized identification method based on SVM (Support Vector Machine) and GMM (Gaussian Mixture Model) |
CN103177722A (en) * | 2013-03-08 | 2013-06-26 | 北京理工大学 | Tone-similarity-based song retrieval method |
CN103226948A (en) * | 2013-04-22 | 2013-07-31 | 山东师范大学 | Audio scene recognition method based on acoustic events |
-
2013
- 2013-08-01 CN CN201310332073.6A patent/CN103971702A/en active Pending
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101477798A (en) * | 2009-02-17 | 2009-07-08 | 北京邮电大学 | Method for analyzing and extracting audio data of set scene |
CN101587710A (en) * | 2009-07-02 | 2009-11-25 | 北京理工大学 | A kind of many code books coding parameter quantification method based on the audio emergent event classification |
CN102509545A (en) * | 2011-09-21 | 2012-06-20 | 哈尔滨工业大学 | Real time acoustics event detecting system and method |
CN102799899A (en) * | 2012-06-29 | 2012-11-28 | 北京理工大学 | Special audio event layered and generalized identification method based on SVM (Support Vector Machine) and GMM (Gaussian Mixture Model) |
CN103177722A (en) * | 2013-03-08 | 2013-06-26 | 北京理工大学 | Tone-similarity-based song retrieval method |
CN103226948A (en) * | 2013-04-22 | 2013-07-31 | 山东师范大学 | Audio scene recognition method based on acoustic events |
Non-Patent Citations (2)
Title |
---|
蒋刚 等: "《工业机器人》", 31 January 2011 * |
韩纪庆 等: "《音频信息检索理论与技术》", 31 March 2011 * |
Cited By (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105679313A (en) * | 2016-04-15 | 2016-06-15 | 福建新恒通智能科技有限公司 | Audio recognition alarm system and method |
CN110800053A (en) * | 2017-06-13 | 2020-02-14 | 米纳特有限公司 | Method and apparatus for obtaining event indications based on audio data |
CN107527617A (en) * | 2017-09-30 | 2017-12-29 | 上海应用技术大学 | Monitoring method, apparatus and system based on voice recognition |
CN107910019A (en) * | 2017-11-30 | 2018-04-13 | 中国科学院微电子研究所 | A kind of human acoustical signal's processing and analysis method |
CN111326172A (en) * | 2018-12-17 | 2020-06-23 | 北京嘀嘀无限科技发展有限公司 | Conflict detection method and device, electronic equipment and readable storage medium |
WO2020140552A1 (en) * | 2018-12-31 | 2020-07-09 | 瑞声声学科技(深圳)有限公司 | Haptic feedback method |
CN110223715A (en) * | 2019-05-07 | 2019-09-10 | 华南理工大学 | It is a kind of based on sound event detection old solitary people man in activity estimation method |
CN110223715B (en) * | 2019-05-07 | 2021-05-25 | 华南理工大学 | Home activity estimation method for solitary old people based on sound event detection |
CN111599379A (en) * | 2020-05-09 | 2020-08-28 | 北京南师信息技术有限公司 | Conflict early warning method, device, equipment, readable storage medium and triage system |
CN111599379B (en) * | 2020-05-09 | 2023-09-29 | 北京南师信息技术有限公司 | Conflict early warning method, device, equipment, readable storage medium and triage system |
CN113670434A (en) * | 2021-06-21 | 2021-11-19 | 深圳供电局有限公司 | Transformer substation equipment sound abnormality identification method and device and computer equipment |
CN113421544A (en) * | 2021-06-30 | 2021-09-21 | 平安科技(深圳)有限公司 | Singing voice synthesis method and device, computer equipment and storage medium |
CN113421544B (en) * | 2021-06-30 | 2024-05-10 | 平安科技(深圳)有限公司 | Singing voice synthesizing method, singing voice synthesizing device, computer equipment and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN103971702A (en) | Sound monitoring method, device and system | |
CN102664006B (en) | Abnormal voice detecting method based on time-domain and frequency-domain analysis | |
CN106504754B (en) | A kind of real-time method for generating captions according to audio output | |
CN104732978B (en) | The relevant method for distinguishing speek person of text based on combined depth study | |
CN109616140B (en) | Abnormal sound analysis system | |
CN103971700A (en) | Voice monitoring method and device | |
CN107527617A (en) | Monitoring method, apparatus and system based on voice recognition | |
CN101494049A (en) | Method for extracting audio characteristic parameter of audio monitoring system | |
CN105841797A (en) | Window motor abnormal noise detection method and apparatus based on MFCC and SVM | |
CN103117061A (en) | Method and device for identifying animals based on voice | |
Schröder et al. | Classification of human cough signals using spectro-temporal Gabor filterbank features | |
CN104732972B (en) | A kind of HMM Application on Voiceprint Recognition based on classified statistics is registered method and system | |
CN108520753A (en) | Voice lie detection method based on the two-way length of convolution memory network in short-term | |
CN105608823A (en) | Optical fiber security and protection method and system based on principal component analysis | |
CN110522462A (en) | The multi-modal intelligent trial system of one kind and method | |
CN115102789A (en) | Anti-communication network fraud studying, judging, early-warning and intercepting comprehensive platform | |
CN103021421A (en) | Multilevel screening detecting recognizing method for shots | |
CN115910097A (en) | Audible signal identification method and system for latent fault of high-voltage circuit breaker | |
CN110931024B (en) | Audio-based prediction method and system for natural mating result of captive pandas | |
CN104091104B (en) | Multi-format audio perceives the characteristics extraction of Hash certification and authentication method | |
CN109389994A (en) | Identification of sound source method and device for intelligent transportation system | |
CN115854269A (en) | Leakage hole jet flow noise identification method and device, electronic equipment and storage medium | |
CN108470564A (en) | According to the artificial intelligence approach of audio identification personality characteristics | |
CN111524523A (en) | Instrument and equipment state detection system and method based on voiceprint recognition technology | |
CN106297805A (en) | A kind of method for distinguishing speek person based on respiratory characteristic |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20140806 |