CN109727603B - Voice processing method and device, user equipment and storage medium - Google Patents

Voice processing method and device, user equipment and storage medium Download PDF

Info

Publication number
CN109727603B
CN109727603B CN201811467944.4A CN201811467944A CN109727603B CN 109727603 B CN109727603 B CN 109727603B CN 201811467944 A CN201811467944 A CN 201811467944A CN 109727603 B CN109727603 B CN 109727603B
Authority
CN
China
Prior art keywords
decoding
decoding path
acoustic
score
bifurcation point
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811467944.4A
Other languages
Chinese (zh)
Other versions
CN109727603A (en
Inventor
邵俊尧
钱胜
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Baidu Netcom Science and Technology Co Ltd
Original Assignee
Beijing Baidu Netcom Science and Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Baidu Netcom Science and Technology Co Ltd filed Critical Beijing Baidu Netcom Science and Technology Co Ltd
Priority to CN201811467944.4A priority Critical patent/CN109727603B/en
Publication of CN109727603A publication Critical patent/CN109727603A/en
Application granted granted Critical
Publication of CN109727603B publication Critical patent/CN109727603B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Landscapes

  • Error Detection And Correction (AREA)
  • Machine Translation (AREA)

Abstract

The invention relates to the technical field of information processing, and discloses a voice processing method, a voice processing device, user equipment and a storage medium, which solve the problem that in the decoding process, only the above information is relied on and the below information cannot be effectively utilized. The method comprises the following steps: acquiring voice data; according to the acoustic model, matching to obtain an acoustic score and a decoding path corresponding to each syllable data in the voice data; when the viterbi decodes to a bifurcation point of a decoding path and syllable data after the bifurcation point of the decoding path is matched according to a language model, cutting the decoding path after the bifurcation point according to acoustic scores corresponding to the syllable data after the bifurcation point; matching syllable data on the cut decoding path according to the language model to obtain a language score; and performing viterbi decoding on the voice data frame by frame according to the acoustic score and the language score on the cut decoding path. The embodiment of the invention is suitable for the processing process of voice data.

Description

Voice processing method and device, user equipment and storage medium
Technical Field
The present invention relates to the field of information processing technologies, and in particular, to a voice processing method and apparatus, a user equipment, and a storage medium.
Background
The traditional speech recognition system comprises a decoder, a language model and an acoustic model. After the voice signals are obtained, a plurality of decoding paths are built on a decoding space, the voice signals traverse each decoding path, scoring is carried out on the basis of an acoustic model and a language model, an acoustic model score and a language model score of each decoding path are obtained, an optimal decoding path is determined on the basis of the scores, and a final recognition result is output according to the optimal decoding path.
In the prior viterbi decoding process, a pruning algorithm is usually accompanied, and some paths with lower scores are pruned to accelerate the decoding speed. The mainstream pruning algorithm in the prior art comprises beam pruning and histogram pruning. The beam pruning is the highest score of the total scores of the acoustic scores and the language scores in all the paths after the decoding path branches are determined, and the paths corresponding to the scores with the highest score lower than a preset range are cut. The histogram pruning is to reserve the paths corresponding to the higher scores in the preset number after determining the total scores of the acoustic scores and the language scores in all the paths after the decoding path branches, and to cut the rest paths. In both of the pruning methods, paths which have undergone a large amount of calculation are pruned due to low scores, and the amount of calculation performed before is wasted.
Disclosure of Invention
The invention aims to solve the problem that the prior art only depends on the above information and cannot effectively utilize the following information in the decoding process, and provides a voice processing method, a device, user equipment and a storage medium, so that the following information is better utilized, unnecessary language model scoring is effectively reduced, the decoding speed is improved, and the voice data processing efficiency is improved.
In order to achieve the above object, an embodiment of a first aspect of the present invention provides a speech processing method, where the method includes: acquiring voice data; according to a pre-established acoustic model, matching to obtain an acoustic score corresponding to each syllable data in the voice data and a decoding path where the acoustic score is located; when viterbi decodes a bifurcation point of the decoding path and matches syllable data after the bifurcation point of the decoding path according to a pre-established language model, cutting the decoding path after the bifurcation point according to an acoustic score corresponding to the syllable data after the bifurcation point; matching syllable data on the cut decoding path according to the language model to obtain a language score; and carrying out viterbi decoding on the voice data frame by frame according to the acoustic score and the language score on the cut decoding path.
Further, the clipping the decoding path after the bifurcation point according to the acoustic score corresponding to the syllable data after the bifurcation point includes: and cutting the decoding path after the bifurcation point according to the acoustic scores corresponding to the first syllable data on all the decoding paths after the bifurcation point.
Further, the clipping the decoding path after the bifurcation point according to the acoustic scores corresponding to the first syllable data on all the decoding paths after the bifurcation point includes: comparing the acoustic scores corresponding to the first syllable data on all decoding paths after the bifurcation point to determine the highest acoustic score; and cutting a decoding path where the first syllable data corresponding to the acoustic score in the preset range lower than the highest acoustic score is located.
Further, the performing viterbi decoding on the voice data frame by frame according to the acoustic score and the linguistic score on the clipped decoding path includes: and selecting an optimal decoding path according to the acoustic score and the language score on the cut decoding path, and outputting the optimal decoding path as the recognition result of the voice data.
An embodiment of a second aspect of the present invention provides a speech processing apparatus, including: an acquisition unit configured to acquire voice data; the acoustic matching unit is used for matching to obtain an acoustic score corresponding to each syllable data in the voice data and a decoding path where the acoustic score corresponds to each syllable data according to a pre-established acoustic model; the cutting unit is used for cutting the decoding path behind the bifurcation point according to the acoustic score corresponding to the syllable data behind the bifurcation point when the viterbi decodes to the bifurcation point of the decoding path and the syllable data behind the bifurcation point of the decoding path is matched according to a pre-established language model; the language matching unit is used for matching syllable data on the cut decoding path according to the language model to obtain a language score; and the decoding unit is used for carrying out viterbi decoding on the voice data frame by frame according to the acoustic score and the language score on the cut decoding path.
Further, the clipping unit is further configured to clip the decoding path after the bifurcation point according to the acoustic scores corresponding to the first syllable data on all the decoding paths after the bifurcation point.
Further, the clipping unit is further configured to compare acoustic scores corresponding to the first syllable data on all decoding paths after the bifurcation point, and determine a highest acoustic score; and cutting a decoding path where the first syllable data corresponding to the acoustic score in the preset range lower than the highest acoustic score is located.
Further, the decoding unit is further configured to select an optimal decoding path according to the acoustic score and the linguistic score on the clipped decoding path, and output the optimal decoding path as the recognition result of the voice data.
A third embodiment of the present invention provides a user equipment, which includes a microphone, a processor, a memory, and a computer program stored in the memory and executable on the processor, wherein the microphone is used for acquiring a voice signal, and the processor implements the voice processing method as described above when executing the program.
A fourth aspect of the present invention provides a storage medium having stored therein instructions that, when run on a computer, cause the computer to perform the speech processing method as described above.
By the technical scheme, in the viterbi decoding process, the acoustic scores in the following information are used for pruning in advance, so that unnecessary scoring of the language model is effectively reduced, the decoding speed is increased, and the voice data processing efficiency is improved.
Drawings
Fig. 1 is a schematic flow chart of a speech processing method according to an embodiment of the present invention;
FIG. 2 is a schematic diagram of a decoding network in accordance with one embodiment of the present invention;
FIG. 3 is a schematic diagram of another decoding network in accordance with one embodiment of the present invention;
fig. 4 is a schematic structural diagram of a speech processing apparatus according to an embodiment of the present invention.
Detailed Description
The following detailed description of embodiments of the invention refers to the accompanying drawings. It should be understood that the detailed description and specific examples, while indicating the present invention, are given by way of illustration and explanation only, not limitation.
Fig. 1 is a flowchart illustrating a speech processing method according to an embodiment of the present invention. As shown in fig. 1, the method comprises the steps of:
step 101, acquiring voice data;
step 102, according to a pre-established acoustic model, matching to obtain an acoustic score corresponding to each syllable data in the voice data and a decoding path where the acoustic score is located;
103, when the viterbi decodes to a bifurcation point of a decoding path and syllable data after the bifurcation point of the decoding path is matched according to a pre-established language model, cutting the decoding path after the bifurcation point according to an acoustic score corresponding to the syllable data after the bifurcation point;
104, matching syllable data on the cut decoding path according to the language model to obtain a language score; and
and 105, performing viterbi decoding on the voice data frame by frame according to the acoustic score and the language score on the cut decoding path.
In an embodiment of the present invention, after the received voice signal is subjected to preprocessing, such as front-end processing, feature extraction, and the like, voice data corresponding to the received voice signal is obtained, so that matching between an acoustic model and a language model is performed on the voice data, and a recognition result of the voice data is finally obtained.
The existing front-end processing means that before feature extraction, a received voice signal is generally processed first, so that influences caused by noise and different speakers are eliminated as much as possible, and the processed signal can reflect essential features of voice better. The most common front-end processes in existence are endpoint detection and speech enhancement. The endpoint detection is to distinguish the speech signal from the non-speech signal in the speech signal to accurately determine the starting point of the speech signal. After the endpoint detection, the subsequent processing can be carried out on the voice signal only, which plays an important role in improving the accuracy of the model and the recognition accuracy. The main task of speech enhancement is to eliminate the effect of ambient noise on speech.
The preprocessing techniques such as front-end processing utilized in the embodiments of the present invention may be existing or may appear in the future, and the present invention is not limited thereto.
In one embodiment of the present invention, the pre-established acoustic model is trained based on a CTC (connection temporal Classification) technique. Specifically, feature extraction may be performed on a large amount of speech data to obtain a feature vector of each speech data. And then adding blank labels in the feature vectors at intervals of a preset number of pronunciation units, training the voice data added with the blank labels based on connection time sequence classification, and establishing an acoustic model. Wherein, the acoustic model comprises a plurality of syllable data.
The plurality of syllable data in the acoustic model and the jump relationship between the syllable data can form a large number of decoding paths, and the decoding paths can form a decoding network. For example, fig. 2 is a schematic diagram of a decoding network in accordance with one embodiment of the present invention. As shown in fig. 2, wherein circles represent syllable data in the decoding network and arrows represent jump relationships between syllable data. As can be seen from fig. 2, there are multiple decoding paths in the decoding network. Each decoding path is one possible decoding result when decoding speech data. In the embodiment of the present invention, the process of performing viterbi decoding on the voice data is a process of selecting an optimal decoding path from a plurality of decoding paths in a decoding network according to a feature vector frame of the voice data.
In one embodiment of the invention, the contextual information is leveraged to reduce unnecessary scoring of language models. The specific implementation process is that after the acoustic score and the decoding path corresponding to each syllable data in the acquired voice data are obtained by matching according to the acoustic model, when the syllable data of the decoding path are matched according to the pre-established language model, when the viterbi decodes the bifurcation point of the decoding path, the decoding path after the bifurcation point is cut according to the acoustic score corresponding to the syllable data after the bifurcation point. Specifically, when the viterbi decoder reaches the bifurcation point of the decoding path, the scores of the language models on the following paths are not calculated, but the acoustic scores corresponding to the first syllable data on all paths after the bifurcation point are compared, and the highest acoustic score is determined. And cutting off the decoding path where the first syllable data corresponding to the acoustic score which is lower than 5% of the highest acoustic score is located according to the preset range, for example, the preset range is 5%, so as to obtain the cut decoding path. The setting form of the preset range is not limited in the present invention, and may be the percentage of the above example, or a numerical range, or other forms, and the preset range may be set according to specific needs.
For example, fig. 3 is a schematic diagram of a decoding network according to an embodiment of the present invention. As shown in fig. 3, wherein when the viterbi decodes to reach the bifurcation point a of the decoding path, the acoustic scores of da4 after the point a and ren2 are compared first, thereby determining the highest acoustic score. If the acoustic score of da4 is the highest, the acoustic score of da4 is determined as the highest acoustic score, and if the acoustic score of ren2 is lower than the preset range of the acoustic score of da4, the decoding path corresponding to ren2 is cut out, namely, the path which branches from the point A to the 'Beijing person' and is decoded is directly discarded. Similarly, if the acoustic score of ren2 is the highest acoustic score and the acoustic score of da4 is lower than the preset range of the acoustic score of ren2, the decoding path corresponding to da4 is cut off, and the path branching from point a to "beijing university" is abandoned. In addition, if the acoustic score of da4 is not below the preset range of acoustic scores of ren2, the decoding path corresponding to da4 may be preserved.
By the method for pre-clipping by using the following information, namely when the syllable data after the bifurcation point is subjected to language model scoring, the decoding path is pre-clipped according to the acoustic score on the decoding path after the bifurcation point, and then the language model scoring is carried out on the clipped decoding path, so that unnecessary scoring of the language model is reduced (corresponding to the condition that the decoding path with low acoustic score is clipped), the decoding accuracy is also ensured (corresponding to the condition that the decoding path with higher acoustic score is reserved), and the viterbi decoding speed is improved.
In an embodiment of the present invention, after the clipping, the syllable data on the clipped decoding path may be matched according to a pre-established language model to obtain a language score. The language model applied in the embodiment of the invention is an N-gram language model.
And then, selecting an optimal decoding path according to the acoustic score and the language score on the cut decoding path, and outputting the optimal decoding path as a recognition result of the voice data. Wherein, the pruning technique in the prior art can also be utilized to select the optimal decoding path from the clipped decoding path, which is not described herein again.
According to the embodiment of the invention, in the viterbi decoding process, the acoustic scores in the following information are used for carrying out pruning in advance, so that unnecessary scoring of the language model is effectively reduced, the decoding speed is increased, and the voice data processing efficiency is improved.
Correspondingly, fig. 4 is a schematic structural diagram of a speech processing apparatus according to an embodiment of the present invention. As shown in fig. 4, the apparatus includes an acquisition unit 41 for acquiring voice data; the acoustic matching unit 42 is configured to match an acoustic score corresponding to each syllable data in the speech data and a decoding path where the syllable data is located according to a pre-established acoustic model; a clipping unit 43, configured to, when the viterbi decodes to a bifurcation point of the decoding path and matches the syllable data after the bifurcation point of the decoding path according to a pre-established language model, clip the decoding path after the bifurcation point according to an acoustic score corresponding to the syllable data after the bifurcation point; the language matching unit 44 is used for matching the syllable data on the cut decoding path according to the language model to obtain a language score; and a decoding unit 45, configured to perform viterbi decoding on the speech data frame by frame according to the acoustic score and the language score on the clipped decoding path.
Further, the clipping unit is further configured to clip the decoding path after the bifurcation point according to the acoustic scores corresponding to the first syllable data on all the decoding paths after the bifurcation point. The clipping unit is further configured to compare acoustic scores corresponding to the first syllable data on all decoding paths after the bifurcation point, and determine a highest acoustic score; and cutting a decoding path where the first syllable data corresponding to the acoustic score in the preset range lower than the highest acoustic score is located.
Further, the decoding unit is further configured to select an optimal decoding path according to the acoustic score and the linguistic score on the clipped decoding path, and output the optimal decoding path as the recognition result of the voice data.
Correspondingly, the embodiment of the present invention further provides a user device, which includes a microphone, a processor, a memory, and a computer program stored in the memory and operable on the processor, where the microphone is used to acquire a voice signal, and the processor implements the voice processing method as described above when executing the program.
Accordingly, an embodiment of the present invention further provides a storage medium, where instructions are stored, and when the storage medium runs on a computer, the storage medium causes the computer to execute the voice processing method described above.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to embodiments of the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
In a typical configuration, a computing device includes one or more processors (CPUs), input/output interfaces, network interfaces, and memory.
The memory may include forms of volatile memory in a computer readable medium, Random Access Memory (RAM) and/or non-volatile memory, such as Read Only Memory (ROM) or flash memory (flash RAM). The memory is an example of a computer-readable medium.
Computer-readable media, including both non-transitory and non-transitory, removable and non-removable media, may implement information storage by any method or technology. The information may be computer readable instructions, data structures, modules of a program, or other data. Examples of computer storage media include, but are not limited to, phase change memory (PRAM), Static Random Access Memory (SRAM), Dynamic Random Access Memory (DRAM), other types of Random Access Memory (RAM), Read Only Memory (ROM), Electrically Erasable Programmable Read Only Memory (EEPROM), flash memory or other memory technology, compact disc read only memory (CD-ROM), Digital Versatile Discs (DVD) or other optical storage, magnetic cassettes, magnetic tape magnetic disk storage or other magnetic storage devices, or any other non-transmission medium that can be used to store information that can be accessed by a computing device. As defined herein, a computer readable medium does not include a transitory computer readable medium such as a modulated data signal and a carrier wave.
It should also be noted that the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in the process, method, article, or apparatus that comprises the element.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The above are merely examples of the present application and are not intended to limit the present application. Various modifications and changes may occur to those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present application should be included in the scope of the claims of the present application.

Claims (10)

1. A method of speech processing, the method comprising:
acquiring voice data;
according to a pre-established acoustic model, matching to obtain an acoustic score corresponding to each syllable data in the voice data and a decoding path where the acoustic score is located;
when viterbi decodes a bifurcation point of the decoding path and matches syllable data after the bifurcation point of the decoding path according to a pre-established language model, cutting the decoding path after the bifurcation point according to an acoustic score corresponding to the syllable data after the bifurcation point;
matching syllable data on the cut decoding path according to the language model to obtain a language score; and
and carrying out viterbi decoding on the voice data frame by frame according to the acoustic score and the language score on the cut decoding path.
2. The method of claim 1, wherein the clipping the decoding path after the bifurcation point according to the acoustic score corresponding to the syllable data after the bifurcation point comprises:
and cutting the decoding path after the bifurcation point according to the acoustic scores corresponding to the first syllable data on all the decoding paths after the bifurcation point.
3. The method of claim 2, wherein the clipping the decoding path after the bifurcation point according to the acoustic scores corresponding to the first syllable data on all decoding paths after the bifurcation point comprises:
comparing the acoustic scores corresponding to the first syllable data on all decoding paths after the bifurcation point to determine the highest acoustic score; and
and cutting a decoding path where the first syllable data corresponding to the acoustic score in the preset range lower than the highest acoustic score is located.
4. The method of claim 1, wherein said performing viterbi decoding on said speech data frame by frame according to the acoustic score and the linguistic score on the clipped decoding path comprises:
and selecting an optimal decoding path according to the acoustic score and the language score on the cut decoding path, and outputting the optimal decoding path as the recognition result of the voice data.
5. A speech processing apparatus, characterized in that the apparatus comprises:
an acquisition unit configured to acquire voice data;
the acoustic matching unit is used for matching to obtain an acoustic score corresponding to each syllable data in the voice data and a decoding path where the acoustic score corresponds to each syllable data according to a pre-established acoustic model;
the cutting unit is used for cutting the decoding path behind the bifurcation point according to the acoustic score corresponding to the syllable data behind the bifurcation point when the viterbi decodes to the bifurcation point of the decoding path and the syllable data behind the bifurcation point of the decoding path is matched according to a pre-established language model;
the language matching unit is used for matching syllable data on the cut decoding path according to the language model to obtain a language score; and
and the decoding unit is used for carrying out viterbi decoding on the voice data frame by frame according to the acoustic score and the language score on the cut decoding path.
6. The apparatus of claim 5, wherein the clipping unit is further configured to clip the decoding path after the bifurcation point according to the acoustic scores corresponding to the first syllable data on all decoding paths after the bifurcation point.
7. The apparatus of claim 6, wherein the clipping unit is further configured to compare acoustic scores corresponding to the first syllable data on all decoding paths after the bifurcation point to determine a highest acoustic score; and cutting a decoding path where the first syllable data corresponding to the acoustic score in the preset range lower than the highest acoustic score is located.
8. The apparatus of claim 5, wherein the decoding unit is further configured to select an optimal decoding path according to the acoustic score and the linguistic score on the clipped decoding path, and output the optimal decoding path as the recognition result of the speech data.
9. User equipment comprising a microphone, a processor, a memory and a computer program stored on the memory and executable on the processor, characterized in that the microphone is adapted to acquire speech signals and the processor implements the speech processing method according to any of the claims 1-4 when executing the program.
10. A storage medium having stored therein instructions which, when run on a computer, cause the computer to execute the speech processing method according to any one of claims 1-4.
CN201811467944.4A 2018-12-03 2018-12-03 Voice processing method and device, user equipment and storage medium Active CN109727603B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811467944.4A CN109727603B (en) 2018-12-03 2018-12-03 Voice processing method and device, user equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811467944.4A CN109727603B (en) 2018-12-03 2018-12-03 Voice processing method and device, user equipment and storage medium

Publications (2)

Publication Number Publication Date
CN109727603A CN109727603A (en) 2019-05-07
CN109727603B true CN109727603B (en) 2020-11-03

Family

ID=66295851

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811467944.4A Active CN109727603B (en) 2018-12-03 2018-12-03 Voice processing method and device, user equipment and storage medium

Country Status (1)

Country Link
CN (1) CN109727603B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110956954B (en) * 2019-11-29 2020-12-11 百度在线网络技术(北京)有限公司 Speech recognition model training method and device and electronic equipment
CN113903340A (en) * 2020-06-18 2022-01-07 北京声智科技有限公司 Sample screening method and electronic device
CN113066489A (en) * 2021-03-16 2021-07-02 深圳地平线机器人科技有限公司 Voice interaction method and device, computer readable storage medium and electronic equipment

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103871403A (en) * 2012-12-13 2014-06-18 北京百度网讯科技有限公司 Method of setting up speech recognition model, speech recognition method and corresponding device

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101030369B (en) * 2007-03-30 2011-06-29 清华大学 Built-in speech discriminating method based on sub-word hidden Markov model
CN102376305B (en) * 2011-11-29 2013-06-19 安徽科大讯飞信息科技股份有限公司 Speech recognition method and system
WO2014025682A2 (en) * 2012-08-07 2014-02-13 Interactive Intelligence, Inc. Method and system for acoustic data selection for training the parameters of an acoustic model
CN105513589B (en) * 2015-12-18 2020-04-28 百度在线网络技术(北京)有限公司 Speech recognition method and device
CN105845128B (en) * 2016-04-06 2020-01-03 中国科学技术大学 Voice recognition efficiency optimization method based on dynamic pruning beam width prediction
CN107403620A (en) * 2017-08-16 2017-11-28 广东海翔教育科技有限公司 A kind of audio recognition method and device
CN108510990A (en) * 2018-07-04 2018-09-07 百度在线网络技术(北京)有限公司 Audio recognition method, device, user equipment and storage medium

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103871403A (en) * 2012-12-13 2014-06-18 北京百度网讯科技有限公司 Method of setting up speech recognition model, speech recognition method and corresponding device

Also Published As

Publication number Publication date
CN109727603A (en) 2019-05-07

Similar Documents

Publication Publication Date Title
US9875739B2 (en) Speaker separation in diarization
CN109473123B (en) Voice activity detection method and device
WO2021128741A1 (en) Voice emotion fluctuation analysis method and apparatus, and computer device and storage medium
CN108630193B (en) Voice recognition method and device
CN109727603B (en) Voice processing method and device, user equipment and storage medium
CN106328127B (en) Speech recognition apparatus, speech recognition method, and electronic device
CN109817198B (en) Speech synthesis method, apparatus and storage medium
US10475484B2 (en) Method and device for processing speech based on artificial intelligence
CN102013253B (en) Speech recognition method based on speed difference of voice unit and system thereof
CN111797632B (en) Information processing method and device and electronic equipment
US20150179166A1 (en) Decoder, decoding method, and computer program product
CN106504768A (en) Phone testing audio frequency classification method and device based on artificial intelligence
CN110060665A (en) Word speed detection method and device, readable storage medium storing program for executing
CN111613215B (en) Voice recognition method and device
CN113053390B (en) Text processing method and device based on voice recognition, electronic equipment and medium
CN112750441B (en) Voiceprint recognition method and device, electronic equipment and storage medium
CN107680584B (en) Method and device for segmenting audio
CN111883113A (en) Voice recognition method and device
CN112259084B (en) Speech recognition method, device and storage medium
CN111292725B (en) Voice decoding method and device
CN106653029A (en) Audio batch segmentation method and device
CN109559733B (en) Voice rhythm processing method and device
CN108922556B (en) Sound processing method, device and equipment
CN114420136A (en) Method and device for training voiceprint recognition model and storage medium
JP5786717B2 (en) Data processing apparatus, computer program thereof, and data processing method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant