EP0614075A2 - Method for speech coding using Trellis Coded Quantization for Linear Predictive Coding quantization - Google Patents
Method for speech coding using Trellis Coded Quantization for Linear Predictive Coding quantization Download PDFInfo
- Publication number
- EP0614075A2 EP0614075A2 EP94103204A EP94103204A EP0614075A2 EP 0614075 A2 EP0614075 A2 EP 0614075A2 EP 94103204 A EP94103204 A EP 94103204A EP 94103204 A EP94103204 A EP 94103204A EP 0614075 A2 EP0614075 A2 EP 0614075A2
- Authority
- EP
- European Patent Office
- Prior art keywords
- quantization
- state
- lsp
- trellis
- fact
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000013139 quantization Methods 0.000 title claims abstract description 129
- 238000000034 method Methods 0.000 title claims abstract description 58
- 239000013598 vector Substances 0.000 claims description 40
- 230000007704 transition Effects 0.000 claims description 29
- 238000005192 partition Methods 0.000 claims description 24
- 238000012549 training Methods 0.000 claims description 6
- 239000000654 additive Substances 0.000 claims 1
- 230000000996 additive effect Effects 0.000 claims 1
- 238000001228 spectrum Methods 0.000 abstract description 2
- 238000013459 approach Methods 0.000 description 5
- 238000004891 communication Methods 0.000 description 4
- 238000009472 formulation Methods 0.000 description 3
- 239000000203 mixture Substances 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 244000141353 Prunus domestica Species 0.000 description 2
- 230000008901 benefit Effects 0.000 description 2
- 230000014509 gene expression Effects 0.000 description 2
- 230000003595 spectral effect Effects 0.000 description 2
- 241000282336 Nasua Species 0.000 description 1
- 238000005311 autocorrelation function Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- HPNSNYBUADCFDR-UHFFFAOYSA-N chromafenozide Chemical compound CC1=CC(C)=CC(C(=O)N(NC(=O)C=2C(=C3CCCOC3=CC=2)C)C(C)(C)C)=C1 HPNSNYBUADCFDR-UHFFFAOYSA-N 0.000 description 1
- 239000012141 concentrate Substances 0.000 description 1
- 238000005314 correlation function Methods 0.000 description 1
- 230000003247 decreasing effect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000002349 favourable effect Effects 0.000 description 1
- 238000013138 pruning Methods 0.000 description 1
- 238000000638 solvent extraction Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000001755 vocal effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/04—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
- G10L19/06—Determination or coding of the spectral characteristics, e.g. of the short-term prediction coefficients
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L19/00—Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
- G10L19/002—Dynamic bit allocation
Definitions
- the present invention relates to a method for speech coding as set forth in the preamble of claim 1 and a speech coder as set forth in the preamble of claim 17.
- Trellis Coded Modulation From the field of the communication theory, and in particular from the modulation theory the Trellis Coded Modulation technique is well known.
- the Trellis coded Modulation paradigm combined with well-known quantization theories, gave rise to the Trellis Coded Quantization (TCQ) algorithm.
- TCQ Trellis Coded Quantization
- TCQ is a recent technique for efficient scalar encoding of any source.
- the main object of the present invention is therefore substantially an efficient and effective way how to apply the TCQ technique.
- the method for speech coding is constructed as set forth in claim 1 and the speech coder as set forth in claim 17.
- figure 1 is showed a general structure of a Trellis Coded Quantizer
- figure 2 ia a Trellis Coded Quantizer with variable bit allocation
- figure 3 is a TCQ scheme for LSP difference quantization
- figure 4 is an updated paths in the TCQ.
- the trellis encoder is completely specified by:
- the 4-state trellis is fully connected.
- the number associated to each state transition (branch) represents the quantization value corresponding to that branch.
- the trellis is a N-stage one, that is, it is employed for coding a N-component input vector.
- a trellis scheme like the one in Figure 1 is an example of what is generally found in the literature. That is, the topological configuration of the trellis is the same at each quantization step.
- the quantization level number is the same at each quantization step.
- This configuration may be not the ideal one in case one needs to quantize a vector whose scalar components have a different 'importance scale', according to a predefined performance criterium. In this case, a different bit/sample number may be necessary for each vector component.
- both the bit/sample increase and the decrease can be easily realized while carrying out the Viterbi algorithm in the encoding process.
- the addition of parallel transitions implies an increase in the evaluation of the local transition state metrics.
- pruning a state transition branch implies the assignment of a corresponding 'infinite' local transition state metric.
- LSP line spectrum pairs
- the ordering property of the LSP parameters can be exploited by quantizing the differences between adjacent LSP frequencies instead of the absolute values of the LSP frequencies.
- a proper bit allocation can be assigned to each LSP difference according to its perceptual importance.
- the TCQ algorithm When applied to the quantization of the LSP differences, the TCQ algorithm proves itself to be particularly effective, since the quantization error accumulated in quantizing the - say - first ( i -1)-th LSP differences can be taken into account in the search of the optimum quantization level for the i -th LSP difference.
- Each trellis state will be assigned a "history path", at each i -th trellis stage; each state transition belonging to this history path will correspond to a pointer to the quantization level of the corresponding LSP difference.
- the i -th quantized LSP can be reconstructed (note that this reconstructed LSP will be different - in general - for each trellis state).
- the quantity refers to the quantized i -th LSP difference, according to the corresponding quantization level belonging to the j -th state history path.
- transition cost from each j -th state to each possible k -th future state is computed.
- This transition cost is related to the quantization level associated to the corresponding transition branch; with reference to Figure 3, the transition cost is denoted as C jk .
- C jk depends on the quantization error which is measured (according to a proper metric) as function of the "transitional" quantization level.
- the final state with the minimum accumulated cost is selected as the "winning" one. Its index (or, equivalently, the index of the corresponding initial state) is transmitted, together with the state transition labels of its history path.
- the initial winning state index and the state transition labels of its history path are input. All the LSP differences can be recovered from the state transition label pointers to the quantization level table. Afterward, the LSP frequencies can be reconstructed.
- the working principle is analogous to the one described previously for the 1-D (i.e. intra-frame) case, which simply exploits the LSP,ordering property.
- the combination weights are the 2-D predictor coefficients.
- a multi-coefficient 2-D predictor can be employed, both in the intra-frame and in the inter-frame sense, as follows: It is worth to note that the predictor length is not necessarily the same in each dimension.
- f i ( n ) a i f i -1 ( n ) + b i f i ( n -1) + c i f i -1 ( n -1) (4) that is, we introduce another inter-frame/intra-frame dependency, namely the one related to the previous (in the intra-frame sense) LSP of the previous (in the inter-frame sense) frame.
- the third weighting coefficient can be determined in an 'optimal' way, as will be described in a following section.
- the concept can be extended further, by introducing a multi-coefficient multi-dimensional predictor, operating with different prediction orders, according to the prediction 'direction' (i.e. intra-frame, inter-frame, various intra-frame/inter-frame combinations).
- LSP differences can be recovered from the best state information and the related history path.
- the LSP values can then be reconstructed by re-adding the previously (both in the intra-frame and in the inter-frame sense) reconstructed parameters, after weighting them by the corresponding predictor coefficients.
- TCQ of the LSP parameters in a generical differential sense
- TCQ of the LSP parameters can be carried out according to any suitable metric that allows to measure an overall distortion as function of successive local distortions.
- a simple mean squared error could be used as the local metric for the quantization error.
- the transition cost defined previously i.e. see the 1-D case
- the transition cost defined previously could be defined as:
- a weighted MSE could be employed, following (e.g.) the guidelines specified in K.K. Paliwal, B.S. Atal, "Efficient Vector Quantization of LPC Parameters at 24 Bits/Frame", Proc. ICASSP '91, pagg. 661-664, where the spectral content of the speech signal at the LSP frequency location is taken into account explicitly.
- a WMSE criterion that considers the relative weight of the specific LSP that is being quantized could also be considered.
- formula (5) could be re-written as: where f (.,.) is a (one-dimensional or two-dimensional) weighting function that would take into account the differential LSP to be quantized and/or the quantization level that is being considered.
- a recursive structure may be used for the computation of the reflection coefficients, starting either from the values of the autocorrelation function (and thereby using the well-known Leroux-Gueguen algorithm) or from the values of the signal covariance function (by employing the so-called covariance-lattice formulation, as explained in A. Cumani, "On a Covariance-Lattice Algorithm fof Linear Prediction", Proc. ICASSP '82, pagg. 651-654).
- Leroux-Gueguen algorithm should be reformulated properly in order to take into account the eventual quantization of the reflection coefficients after their computation at each step of the recursion.
- each reflection coefficients can be computed as function of each particular state.
- the computed reflection coefficient can be quantized according to the quantization level subset 'seen' by each particular trellis state.
- the recursive algorithm for reflection coefficient computation, with embedded TCQ may be stated as follows (only the formulation related to the covariance-lattice approach is given, since the corresponding formalism for the autocorrelation approach may be derived in an analogous way; also, note that the formalism used resembles the one described in: A. Cumani, "On a Covariance-Lattice Algorithm for Linear Prediction", Proc. ICASSP '82, pagg. 651-654.
- the i-th LSP difference must be quantized; its value is computed by taking the difference between the i-th LSF and the reconstructed (i.e. quantized) ( i - 1)-th LSF.
- This reconstructed ( i -1)-th LSF is different for each trellis state; the i -th LSF difference thus obtained must be quantized accordingly to the level partition "seen" by the corresponding trellis state.
- TCQ training procedure in particular we start from a unique set of quantization values for each state subset; these values can be found by using a standard scalar quantization clustering procedure.
- each possible quantization path is assigned a partition; the corresponding "cluster vector" can be derived by simply taking a proper mean of each partition value and assigning this mean value to the corresponding path state.
- the LSF value training sequence is again input to the TCQ; a new partition set can be generated and the corresponding set of cluster vectors can be found.
- Trellis Coded Vector Quantization is a generalization of the TCQ concept. It has been introduced in T.G. Fisher, M.W. Marcellin, M.Wang, "Trellis Coded Vector Quantization", IEEE Trans. on Information Theory, Vol. IT-37, Nov. 1991. and, again, consists of using a structured codebook with an expanded set of quantization levels.
- the trellis structure prunes the expanded number of quantization reproduction vectors down to the desired encoding rate.
- the TCVQ procedure is carried out in exactly the same way as for the TCQ counterpart, both in the 1 -D case (i.e., taking into account only the intra-frame dependency of the LSP parameters), and in the case of multi-dimensional prediction (i.e., exploiting both the intra-frame and the inter-frame dependency of LSP parameters, with any prediction length in either direction).
- the TCVQ procedure can be carried out by recursive quantization of reflection coefficient couples (if the subvector dimension is actually 2), by using the same strategy employed for the TCQ case.
- TCVQ generalization for the LSP multi-dimensional predictor case and for the reflection coefficients case can be derived in a straightforward manner following the corresponding TCQ descriptions.
- the trellis level reoptimization procedure can be carried out in an analogous way as for the TCQ case.
- the vector clusters can be constructed in an iterative way, as function of the different trellis states and of the corresponding encoding paths. These clusters are obtained as 'centroid' (according to a predetermined metric) of corresponding partitions of the input vector set.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Health & Medical Sciences (AREA)
- Spectroscopy & Molecular Physics (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Compression, Expansion, Code Conversion, And Decoders (AREA)
- Error Detection And Correction (AREA)
- Transmission Systems Not Characterized By The Medium Used For Transmission (AREA)
Abstract
Description
- The present invention relates to a method for speech coding as set forth in the preamble of
claim 1 and a speech coder as set forth in the preamble of claim 17. - In the telecommunication field it is useful to transmit the information both vocal and video using as less numbers of bits as possible without losing part of the information transmitted.
- Such aim is achieved by means of suitable coding techniques.
- There are a lot of these techniques: each of them has its characteristic features. From the field of the communication theory, and in particular from the modulation theory the Trellis Coded Modulation technique is well known. The Trellis coded Modulation paradigm, combined with well-known quantization theories, gave rise to the Trellis Coded Quantization (TCQ) algorithm.
- TCQ is a recent technique for efficient scalar encoding of any source.
- In particular, it has been introduced by M.W. Marcellin, T.G. Fisher, in "Trellis Coded Quantization of Memoryless and Gauss-Markov Sources", IEEE Trans. on Communication, vol. 38, No. 1, January 1990, for encoding memoryless Gauss-Markov sources. The feature of the trellis coded quantization approach is the use of a structured codebook with an expanded set of quantization levels. Based on the notion of set partitioning introduced by G. Ungerboek, in "Channel Coding with Multilevel/Phase Signals", IEEE Trans. on Information Theory, Vol. IT-28, Jan. 1982, the trellis structure then prunes the expanded number of quantization levels down to the desired encoding rate. The encoder uses the Viterbi algorithm for finding a vector of quantized scalars that is closest (according to a predetermined metric) to the unquantized vector.
- Because of the complexity of the matter, only a good knoledge of both the modulation theory (TCM) and of the quantization theory allows a proper implementation and exploitation of this quantization technique.
- The main object of the present invention is therefore substantially an efficient and effective way how to apply the TCQ technique.
- According to the invention, therefore, the method for speech coding is constructed as set forth in
claim 1 and the speech coder as set forth in claim 17. - Further features of the invention are explained in the depending claims.
- Enbondiments of the invention will now be explained in detail with reference to the accompanying drawings, in which figure 1 is showed a general structure of a Trellis Coded Quantizer; figure 2 ia a Trellis Coded Quantizer with variable bit allocation; figure 3 is a TCQ scheme for LSP difference quantization; figure 4 is an updated paths in the TCQ.
- The trellis encoder is completely specified by:
- The trellis topology (i.e. the connections among the successive states or, equivalently, the description of the underlying finite-state machine).
- The quantization levels associated with each state transition.
- In Figure 1, a possible structure for a TCQ is depicted. The labels assigned to each trellis branch are arbitrary, as well as the quantization level partition.
- This particular structure is depicted for the sake of clearness and may not correspond to a 'physical' one.
- The 4-state trellis is fully connected. The number associated to each state transition (branch) represents the quantization value corresponding to that branch. The trellis is a N-stage one, that is, it is employed for coding a N-component input vector.
- Note that 8 possible scalar quantization values are present; however, each trellis state 'sees' only a 4-value subset, as function of its transition to a future state.
- In principle, an exhaustive procedure should be employed for identifying the best quantized vector with respect to a N-value input vector.
- This means that one should identify each possible trellis path and, according to the quantization value associated to each path step, an overall quantization error should be constructed. In the example depicted in Figure 1, this exhaustive procedure would imply identifying 4n quantized vectors and measuring the quantization error for each of them.
- This tremendous amount of computation is avoided by using the well-known Viterbi algorithm as descripted by G.D. Forney, in "The Viterbi algorithm", IEEE Trans. on Information Theory, Vol. IT-28, Jan. 1982, that, although operating in a step-by-step fashion, guarantees to find the optimal solution.
- With reference to Figure 1, it is easily understood' that, in case a scalar quantization technique was employed for encoding each input vector component, 3N bits would be necessary.
- On the contrary, by using a TCQ, the quantization process would require 2N + 2 bits (where the binary representation of the TCQ 'winning' best initial -or final- state is taken into account).
- A trellis scheme like the one in Figure 1 is an example of what is generally found in the literature. That is, the topological configuration of the trellis is the same at each quantization step.
- Similarly, the quantization level number is the same at each quantization step.
- This configuration may be not the ideal one in case one needs to quantize a vector whose scalar components have a different 'importance scale', according to a predefined performance criterium. In this case, a different bit/sample number may be necessary for each vector component.
- This problem can be solved in the following way: suppose to start with a given topological configuration for the trellis (that is, suppose to start with the trellis depicted in Figure 1); an increase-decrease in the bit/sample assignment at each quantization step can be obtained as follows.
- Bit/sample number increase:
this can be obtained by simply adding one or more parallel transitions to the state branches (that is, we have multiple branches at each state transition). The bit/sample number is thus increased according to the number of parallel transitions that are added.
Referring to the example in Figure 2, it is immediate to verify that in the first quantization step (Step 1) we have doubled the quantization level number (from 8 to 16) and, accordingly, the bit/sample configuration (i.e. from 2 to 3 bits).
The quantization level partition associated to each parallel transition can be derived from the optimal set partition theory, as described in M.W. Marcellin, T.G. Fisher, "Trellis Coded Quantization of Memoryless and Gauss-Markov Sources", IEEE Trans. on Communication, vol. 38, No. 1, January 1990, and G. Ungerboek, "Channel Coding with Multilevel/Phase Signals" , IEEE Trans. on Information Theory, Vol. IT-28, Jan. 1982. - Bit/sample number decrease:
This can be obtained by changing the trellis topology in a trivial way.
In particular, the state transition number can be pruned down to the desired encoding rate.
With reference to Figure 2, in the third quantization step (Step 3) the encoding rate is halved (and, obviously, the same is true for the quantization level number) since only a subset of the trellis states can be reached. In the third quantization step, the state transition choice is dichotomic, therefore allowing for a single bit/sample in the quantization of the corresponding vector component. - It is worth to note that from the implementation point of view, both the bit/sample increase and the decrease can be easily realized while carrying out the Viterbi algorithm in the encoding process. In particular, the addition of parallel transitions implies an increase in the evaluation of the local transition state metrics. On the contrary, pruning a state transition branch implies the assignment of a corresponding 'infinite' local transition state metric.
- In recent years line spectrum pairs (LSP) representation of LPC parameters has become popular in speech coding applications. The LSP are frequency domain parameters strictly related to the formants: the position of a pair of frequencies gives the position of the formant, while their difference carries information about the width of the spectral peak.
- The ordering property of the LSP parameters can be exploited by quantizing the differences between adjacent LSP frequencies instead of the absolute values of the LSP frequencies.
- A proper bit allocation can be assigned to each LSP difference according to its perceptual importance.
- When applied to the quantization of the LSP differences, the TCQ algorithm proves itself to be particularly effective, since the quantization error accumulated in quantizing the - say - first (i -1)-th LSP differences can be taken into account in the search of the optimum quantization level for the i-th LSP difference.
- Each trellis state will be assigned a "history path", at each i-th trellis stage; each state transition belonging to this history path will correspond to a pointer to the quantization level of the corresponding LSP difference. By adding all the LSP differences of each state history path up to the i-th trellis stage, the i-th quantized LSP can be reconstructed (note that this reconstructed LSP will be different - in general - for each trellis state).
- As an example, suppose that the i-th LSP difference must be quantized.
- The following operations will be performed:
- For each j-th state of the(i - 1)-th trellis stage, the corresponding (i - 1)-th LSP is reconstructed, by adding all the quantized LSP differences belonging to the j-th state history path.
- For each j-th state, the LSP difference between the input i-th LSP and the reconstructed (i - 1)-th LSP is computed.
- This difference is quantized in with a suitable metric, according to the i-th stage quantizer level partition 'seen' by each j-th state.
- The Viterbi algorithm is then acted upon on each trellis state, in order to determine the best previous state and, therefore, the updated history path. Furthermore, the quantization accumulated cost is updated for each state; in particular, this cost is consistent with the metric used for each LSP difference quantization.
- An example of this procedure is depicted in Figure 3, where a simple 4-state trellis with 4 quantization levels (for each quantization step) and 1 bit/branch is used.
- Suppose that the nd LSP is input. Hence, the corresponding LSP difference must be quantized.
-
- It is clear that by adding all the quantized LSP difference belonging to the generical state history path, the reconstructed st LSP may be obtained, as function of the state under consideration.
-
-
- Afterward, the transition cost from each j-th state to each possible k-th future state is computed. This transition cost is related to the quantization level associated to the corresponding transition branch; with reference to Figure 3, the transition cost is denoted as C jk . In particular, C jk depends on the quantization error which is measured (according to a proper metric) as function of the "transitional" quantization level.
- To be more specific, let L jk the quantization level associated to the transition between the j-th state and the k-th one. We can write:
- Starting from
state 0, compute: - Starting from
state 1, compute: - Similarly, the procedure is repeated for
state 2 andstate 3. - After all the transition costs have been computed, the best state path up to the 3rd trellis stage must be updated (for each trellis state). The well-known Viterbi algorithm is used to this goal. As an example, referring to Figure 3, we have the following:
- Current 'observation' state:
state 0
- Compute the candidate accumulated cost with respect to the previous state 0:state 0 overall cost accumulated so far. - Compute the candidate accumulated cost with respect to the previous state 2:
state 2 overall cost accumulated so far. - If A < B the
state 0 is considered to be the best previous state with respect to thecurrent state 0; thenew state 0 path is determined by concatenating thestate 0 path with the state 0 -state 0 transition.
On the contrary, if A > B, thestate 0 path is updated according to thestate 2 path and to the state 2 -state 0 transition. - The same procedure applies for the 'observation' states 2, 3, 4.
- Best previous state with respect to state 0: state0
- Best previous state with respect to state 1: state2
- Best previous state with respect to state 2: state3
- Best previous state with respect to state 3: state3
- We are ready for the next LSP quantization.
- After the p-th LSP difference (p being the predictor order) has been quantized, the final state with the minimum accumulated cost is selected as the "winning" one. Its index (or, equivalently, the index of the corresponding initial state) is transmitted, together with the state transition labels of its history path.
- At the decoder, the initial winning state index and the state transition labels of its history path are input. All the LSP differences can be recovered from the state transition label pointers to the quantization level table. Afterward, the LSP frequencies can be reconstructed.
- Besides an intra-frame correlation of the LSP parameters (i.e. the ordering property), it is possible to take advantage of the strong inter-frame correlation; this may lead to efficient quantization schemes that can operate, for instance, in a two-dimensional differential fashion. A possible application is described in C.C. Kuo, F.R. Jean, H.C. Wang, "Low Bit-Rate Quantization of LSP Parameters Using Two-Dimensional Differential Coding", Proc. ICASSP '92, pagg. 97-100, where a two-dimensional differential coding scheme is shown to significantly improve the effectiveness of the quantization scheme as function of the desired encoding rate.
- In particular, in C.C. Kuo, F.R. Jean, H.C. Wang, "Low Bit-Rate Quantization of LSP Parameters Using Two-Dimensional Differential Coding", Proc. ICASSP '92, pagg. 97-100, the inter-frame/intra-frame dependency of the i-th LSP at frame n may be expressed as:
where p is the predictor order and a i and b i are the coefficients of an optimal two-dimensional (i.e. 2 - D) predictor; these coefficients can be estimated from a long sequence of speech, as described in C.C. Kuo, F.R. Jean, H.C. Wang, "Low Bit-Rate Quantization of LSP Parameters Using Two-Dimensional Differential Coding", Proc. ICASSP '92, pagg. 97-100, f i (n) is the current LSP estimation; f i -1(n) and f i (n-1) are previously quantized parameters. - Therefore, it is possible to transmit the difference between the exact value and the estimated value of the current LSP.
- It is clear that this quantization scheme may not be the optimal one in case channel errors occur during the transmission of the LSP difference information.
- It is possible to cope with this problem by employing a careful design of the 2-D predictor coefficients. However; this issue will be described in greater details in a next paragraph. For the time being, we will assume to deal with proper 2-D predictor coefficients and will concentrate on the TCQ functioning in this case.
- The working principle is analogous to the one described previously for the 1-D (i.e. intra-frame) case, which simply exploits the LSP,ordering property. In particular, it is worth to note that the 1-D case can be considered as a particular case of the 2-D scheme, by putting a i = 1 and b i = 0.
-
- For each j-th state, the 2-D LSP difference is computed between the input i-th LSP and the weighted combination of:
- the reconstructed
- the corresponding i-th quantized LSP derived in the previous frame LSP i (n-1).
- The combination weights are the 2-D predictor coefficients.
- The computed difference is then quantized according to the trellis topology and to the quantization level configuration. The quantization procedure is analogous to the one described in the 1-D case.
- Once all the LSP differences have been quantized and the state paths (as well as their accumulated costs) have been updated, the quantized (n) can be derived, as function of each j-th state under consideration. In particular, we have:
- Iterating this way, all the LSP can be quantized, according to the 2-D predictor behaviour.
-
- A further way to exploit the 'spatial-temporal' redundancy of the LSP parameters is to use a 3-D predictor as follows:
that is, we introduce another inter-frame/intra-frame dependency, namely the one related to the previous (in the intra-frame sense) LSP of the previous (in the inter-frame sense) frame. The third weighting coefficient can be determined in an 'optimal' way, as will be described in a following section. - The concept can be extended further, by introducing a multi-coefficient multi-dimensional predictor, operating with different prediction orders, according to the prediction 'direction' (i.e. intra-frame, inter-frame, various intra-frame/inter-frame combinations).
- Irrespectively of the predictor structure, a difference between a LSP and the corresponding estimated one will be quantized, following the trellis search procedure and the Viterbi algorithm described previously.
- At the decoder site, all the LSP differences can be recovered from the best state information and the related history path. The LSP values can then be reconstructed by re-adding the previously (both in the intra-frame and in the inter-frame sense) reconstructed parameters, after weighting them by the corresponding predictor coefficients.
- It is clear that the TCQ of the LSP parameters (in a generical differential sense) can be carried out according to any suitable metric that allows to measure an overall distortion as function of successive local distortions.
- In particular, a simple mean squared error (MSE) could be used as the local metric for the quantization error. In this respect, the transition cost defined previously (i.e. see the 1-D case) could be defined as:
More in general, a weighted MSE could be employed, following (e.g.) the guidelines specified in K.K. Paliwal, B.S. Atal, "Efficient Vector Quantization of LPC Parameters at 24 Bits/Frame", Proc. ICASSP '91, pagg. 661-664, where the spectral content of the speech signal at the LSP frequency location is taken into account explicitly. Or, a WMSE criterion that considers the relative weight of the specific LSP that is being quantized could also be considered. In this case, formula (5) could be re-written as:
where f(.,.) is a (one-dimensional or two-dimensional) weighting function that would take into account the differential LSP to be quantized and/or the quantization level that is being considered. - Although LSP have proved to be an useful representation of the LPC coefficients with respect to quantization effectiveness, the reflection coefficients are also attractive, for some reasons like:
- Easy control of the filter stability.
- No need of complicated arithmetic procedures to convert them into LSP parameters.
- Possibility of implementing the necessary filter structures in lattice forms, with evident advantages for fix-point computation.
- A recursive structure may be used for the computation of the reflection coefficients, starting either from the values of the autocorrelation function (and thereby using the well-known Leroux-Gueguen algorithm) or from the values of the signal covariance function (by employing the so-called covariance-lattice formulation, as explained in A. Cumani, "On a Covariance-Lattice Algorithm fof Linear Prediction", Proc. ICASSP '82, pagg. 651-654).
- In particular, the Leroux-Gueguen algorithm should be reformulated properly in order to take into account the eventual quantization of the reflection coefficients after their computation at each step of the recursion.
- This gives rise to a slightly modified recursive algorithm in which, starting from the autocorrelation values, the reflection coefficients are computed as follows:
- Let f i (n) be the forward residual of the lattice structure j-th stage and b j (n) be the corresponding backward residual. Then, the expression of the j-th stage residuals as function of the (j - 1)-th stage ones is as follows:
K j being the j-th stage reflection coefficients. - Defining the initial conditions:
s(n) being the lattice structure input signal. - Defining also the following autocorrelation and cross-correlation functions:
- The forward residual autocorrelation at the j-th lattice stage may be expressed by means of the following recursive formula:
- Therefore, the optimal value for the reflection coefficient K j is the one that minimizes the forward residual energy R F (0).
- Once the K j values is computed (and, eventually, quantized), the autocorrelation may be updated, as well as the quantities, R j B , R i FB , R i BF (using expressions similar to the one derived for R i F )
- Again, after defining the trellis topology and the quantization level number at each quantization step, each reflection coefficients can be computed as function of each particular state.
- Its value can therefore take into account the quantization error accumulated along each branch of a generical state path.
- Afterward, the computed reflection coefficient can be quantized according to the quantization level subset 'seen' by each particular trellis state.
- In formulas, the recursive algorithm for reflection coefficient computation, with embedded TCQ may be stated as follows (only the formulation related to the covariance-lattice approach is given, since the corresponding formalism for the autocorrelation approach may be derived in an analogous way; also, note that the formalism used resembles the one described in: A. Cumani, "On a Covariance-Lattice Algorithm for Linear Prediction", Proc. ICASSP '82, pagg. 651-654.
- Given a block of N signal samples: s(0), s(1),..., s(N - 1), compute the covariance Φ ik , for i, k = 0,1,...,p (p being the predictor order):
- Set up F⁰ ij ,B⁰ ij , C⁰ ij , for i,j = 0,1,...,p - 1 using formula (14) of the reference mentioned above. Also, set: m = 0
- For each predictor stage m:
- Compute the m-th reflection coefficient as function of the j-th TCQ state:
- Quantize the reflection coefficient just computed according to the quantizer level partition appertaining to the j-th state. In particular, a non-linear transformation (e.g. log-area ratios) can be done prior to quantization. Each quantization level 'seen' by the j-th state will correspond to a particular state branch connecting the j-th state to a future i-th state (according to the trellis topology).
- The optimal local quantization level thus found is related to a local transition cost, between the j-th state and the i-th one.
- Proceed to the next trellis stage and update the accumulated cost of each state (making use of the accumulated cost of the previous trellis stage and of the local metrics just computed). Also, update the partial quantization path for each state. That is, for each i-th trellis state update the 'forward-backward covariance functions' making use of the function values in the previous j-th state and the quantization level of the reflection coefficient that corresponds to the j-th state -- i-th state transition branch.
- Again, the state with the minimum overall accumulated cost is declared as 'winner'. Its value, together with the trellis labels defining its path, determines the quantized reflection coefficient vector.
As for the LSP case, a proper metric should be employed to carry out the quantization process; in particular, a matric that allows to measure an overall distortion as function of successive local distortions (such as a MSE- or WMSE-based metric) can be suitable.
The quantization procedures outlined in the previous paragraphs may not be the optimal ones (with respect to both the LSP and the reflection coefficients).
In particular, it has been stressed the importance of using suitable metrics, that allow the computation of the accumulated cost as sum of successive partial costs.
From the perceptual point of view, it is well known that the most reliable metric for measuring the effectiveness of the LPC parameter quantization is based on the cepstrual coefficients (e.g. see K.K. Paliwal, B.S. Atal, "Efficient Vector Quantization of LPC Parameters at 24 Bits/Frame", Proc. ICASSP '91, pagg. 661-664). In particular, once two set of predictor coefficients (i.e. before and after quantization) are known, one should compute the corresponding cepstrual coefficient sequences and then measure the MSE (namely, the cepstral distance CD).
However, this procedure is not feasible in carrying out the step-by-step quantization (i.e. as a new LPC parameter is available), as described in the previous TCQ procedures.
Therefore, in order to obtain the set of quantized coefficients that guarantee the best perceptual LPC reproduction, the following steps should carried out: - Reconstruct the decoding path in correspondence of each trellis state, therefore obtaining a set of quantized LPC parameter vectors (i.e. either in terms of LSP or in terms of reflection coefficients).
- For each vector, obtain the corresponding representation in terms of LPC cepstrual coefficients and measure the CD with respect to the cepstral coefficient representation of the unquantized model.
- The trellis parameters to be transmitted should be the ones that define the best LPC vector (in terms of cepstral distance).
- Obtaining the cepstral coefficients from a set of LPC parameters (i.e. LSP or reflection coefficients) is not a trivial task. Therefore, the outlined procedure is likely to be very time-consuming. It is possible to reduce the computation load by reconstructing the decoding path in correspondence of only a subset of the overall trellis states.
- This implies that an implicit assumption is made: namely, the trellis state subset with lowest overall accumulated cost is likely to contain the best trellis state, in terms of CD.
- For better quantization efficiency, the quantization levels are different for each trellis state; the following example clarifies this concept (assume that we are dealing with the quantization of the LSP parameters.
- The same rationales apply for the quantization of the reflection coefficients as well).
- Suppose that the i-th LSP difference must be quantized; its value is computed by taking the difference between the i-th LSF and the reconstructed (i.e. quantized) (i - 1)-th LSF. This reconstructed (i -1)-th LSF is different for each trellis state; the i-th LSF difference thus obtained must be quantized accordingly to the level partition "seen" by the corresponding trellis state.
- Assume that two generical trellis states point to the same subset of quantization level; in standard TCQ procedures (e.g. see M.W. Marcellin, T.G. Fisher, "Trellis Coded Quantization of Memoryless and Gauss-Markov Sources", IEEE Trans. on Communication, vol. 38, No. 1, January 1990) the subset quantization values are the same for the two states; they are only addressed in a different way.
- On the contrary we use different values in the same quantization level subset, as function of TCQ state under consideration.
- In order to obtain this, a proper TCQ training procedure can be adopted; in particular we start from a unique set of quantization values for each state subset; these values can be found by using a standard scalar quantization clustering procedure.
- Afterwards an iterative procedure is adopted in which a long training sequence of LSP is input to the TCQ and the input LSP vector is then assigned to the "partition" corresponding to the obtained TCQ path.
- At the end of the training procedure each possible quantization path is assigned a partition; the corresponding "cluster vector" can be derived by simply taking a proper mean of each partition value and assigning this mean value to the corresponding path state.
- Next, the LSF value training sequence is again input to the TCQ; a new partition set can be generated and the corresponding set of cluster vectors can be found.
- More in details, during a generical iteration step, the following operations can be performed:
- Reset all the partitions belonging to each trellis state. Note that the number of partitions appertaining to each trellis state is equal to the number of quantization levels that can be 'reached' from the state.
- For each LSP input vector:
- Find the optimal quantized vector according to a predefined metric. The quantized vector is identified by specifying the winning starting state and the branch labels along the state path.
- Assign each j-th element of the input vector to a partition whose index corresponds to the branch label of the j-th quantization step along the winning state path. In particular, if the simple MSE is adopted in the quantization phase, the j-th input vector element is simply added to the previous partition value.
- At this point, all the partitions for each trellis state and for each quantization step have been constructed. Each partition 'centroid' can be recomputed by taking an appropriate mean as function of the accumulated partition value and of the number of elements inside the partition. In particular, if the MSE metric is adopted in the quantization phase, each centroid can be computed by taking the simple arithmetic mean of the partition accumulated value.
- Iterating this way it can be observed that the quantization error (in a MSE sense, or in a WMSE sense, according to the metric adopted) is decreasing; although this may not correspond to a performance increase in the cepstral distance sense, it is possible to run the iterative procedure for a fixed iteration number and then choose the quantization level set that guarantees the best performance in terms of cepstral distance.
- Finally, it is worth to note that this iterative reoptimization of the quantization levels is independent on the trellis topology as well as on the LPC parameters under consideration (i.e. LSP or reflection coefficients). However, care must be taken in the determination of the partition centroid, according to the metric used in the quantization phase.
- Trellis Coded Vector Quantization (TCVQ) is a generalization of the TCQ concept. It has been introduced in T.G. Fisher, M.W. Marcellin, M.Wang, "Trellis Coded Vector Quantization", IEEE Trans. on Information Theory, Vol. IT-37, Nov. 1991. and, again, consists of using a structured codebook with an expanded set of quantization levels.
- In particular, instead of dealing with scalar quantization levels, we have an expanded set of reproduction vectors. Again, the trellis structure prunes the expanded number of quantization reproduction vectors down to the desired encoding rate.
- When applied to the LPC parameter quantization, the same strategies can be employed, whether we use the representation in terms of LSP or in terms of reflection coefficients.
- It is clear that, for typical predictor orders (i.e. 10), it is not worth to use high-dimension vectors, in order to maintain a favourable trade-off between performance and encoding rate.
- To be more specific, let's consider the following example:
- Predictor order = 10 (i.e. 10 LPC coefficients to be quantized)
- Scalar quantization versus TCQ case
- Using 3 bits/coefficient (8 quantization levels), scalar quantization implies quantizing the LPC information with 30 bits.
- Using 2 bits/coefficients (8 quantization levels) and a 16-state trellis (which is a good compromise between performance and computation load), TCQ implies quantizing the LPC information with 4 + 20 = 24 bits
- Vector quantization versus TCVQ case
- Dividing the LPC vector in two subvectors of 5 coefficients each and quantizing each subvector with a 2₁₅-element codebook (in order to maintain the same encoding rate as for the scalar quantization case), and using, again, a 16-state trellis, the TCVQ approach allows for an overall encoding rate of 4+14+14 = 32 bits for the LPC information (note that by using simple VQ, we would obtain an encoding rate of 30 bits).
- Dividing the LPC vector in 5 subvectors of 2 coefficients each and quantizing each subvector with a 2₆-element codebook (thus obtaining again an encoding rate of 30 bits for the VQ case), the TCVQ approach with a 16-state trellis would allow to obtain 4+5*5 = 29 bits.
- Actually, the TCVQ technique acted upon subvectors of coefficient couples seems a good compromise between encoding rate and performance.
- The TCVQ procedure is carried out in exactly the same way as for the TCQ counterpart, both in the 1 -D case (i.e., taking into account only the intra-frame dependency of the LSP parameters), and in the case of multi-dimensional prediction (i.e., exploiting both the intra-frame and the inter-frame dependency of LSP parameters, with any prediction length in either direction).
- Besides, when considering the reflection coefficient case (where the prediction-based solution may not be the optimal one), the TCVQ procedure can be carried out by recursive quantization of reflection coefficient couples (if the subvector dimension is actually 2), by using the same strategy employed for the TCQ case.
- To be more specific, a brief description of the TCVQ procedure, when applied to the LSP in the 1-D case, is as follows (assuming to deal with subvectors of coefficient couples. Also, suppose we want to quantize the successive LSP differences): As an example, suppose that the i-th LSP and the (i + 1)-th one must be quantized.
- The following operations will be performed:
- For each j-th state of the (i - 1)-th trellis stage, the corresponding (i - 1)-th LSP is reconstructed, by adding all the quantized LSP difference couples belonging to the j-th state history path.
- For each j-th state, the LSP difference between the input i-th LSP and the reconstructed (i - 1)-th LSP is computed. Furthermore, the LSP difference between the two input LSP parameters can be computed. This gives rise to a 2-component LSP difference vector to be quantized.
- This difference vector is quantized with a suitable metric, according to the i-th stage quantizer level partition 'seen' by each j-th state.
- The Viterbi algorithm is then acted upon on each trellis state, in order to determine the best previous state and, therefore, the updated history path. Furthermore, the quantization accumulated cost is updated for each state; in particular, this coatis consistent with the metric use for each LSP difference quantization.
- Note that the TCVQ generalization for the LSP multi-dimensional predictor case and for the reflection coefficients case can be derived in a straightforward manner following the corresponding TCQ descriptions.
- Finally, also the trellis level reoptimization procedure can be carried out in an analogous way as for the TCQ case. In particular, the vector clusters can be constructed in an iterative way, as function of the different trellis states and of the corresponding encoding paths. These clusters are obtained as 'centroid' (according to a predetermined metric) of corresponding partitions of the input vector set.
functions' that, once determined making use of formulas (12a,b,c) of A. Cumani, "On a Covariance-Lattice Algorithm for Linear Prediction", Proc. ICASSP '82, pagg. 651-654, take into account the quantization values of the previous reflection coefficients along the j-th TCQ state path.
Claims (19)
- Method for speech coding comprising the steps of: receiving in input a sequence of LPC filter coefficients, quantizing said LPC filter coefficients, characterized by said quantization is made using TCQ technique.
- Method according to claim 1 characterized by a variable bit allocation at each quantization step.
- Method according to claim 2 characterized by the fact that a bit rate increase is obtained by adding one or more parallel transitions to the state branches, given a certain trellis topology.
- Method according to claim 2 characterized by the fact that a bit rate decrease is obtained by deleting one or more state branches, given a certain trellis topology.
- Method according to claim 1 characterized by the fact that of each quantization step the quantization error accumulated in quantizing the previous steps can be monitored and, eventually compensed.
- Method according to claim 1 characterized by said LPC filter coefficients are the LSP parameters.
- Method according to claims 5 and 6 characterized by the fact that at each quantization step each trellis state is assigned a history path, each path branch correspond to a pointer to the quantization level of the corresponding LSP value.
- Method according to claims 7 characterized by the fact the history path contains the information associated to each quantization step.
- Method according to claim 6 characterized by the fact that intraframe correlation can be exploited in quantizing the LSP parameters, by means of one dimensional differential prediction schemes along the frequency direction.
- Method according to claim 6 characterized by the fact that interframe correlation can be exploited in quantizing the LSP parameters, by means of one dimensional differential prediction schemes along the time direction.
- Method according to claim 6 characterized by the fact that both interframe correlation and intraframe correlation can be exploited in quantizing the LSP parameters, by means of multi dimensional differential prediction schemes.
- Method according to claim 1 characterized by said LPC filter coefficients are the RC parameters.
- Method according to claims 12 and 5 characterized by coprising the steps of defining trellis topology, defining quantization level number at each quantization step, and computing each RC as function of each particular state.
- Method according to claim 1 characterized by the fact that the quantization error is computed using a metric that has the property of being additive at each quantization step.
- Method according to claim 1 and 14 characterized by comprising the steps of reconstructing the encoding path in correspondence of each trellis state, obtaining a set of quantized LPC parameter vectors, obtaining for each vector, a corresponding representation in terms of LPC cepstral coefficients, measuring the cepstrual distance with respect to the cepstral coefficient representation of the unquantized model, choosing the trellis parameters that define the LPC vector with optimal cepstral distance.
- Method according to claim 1 characterized by adopting a proper TCQ training procedure for a quantization level reoptimization.
- Method according to claim 16 characterized by comprising the step of starting from a set of quantization values for each subset, adopting a iterative procedure with a long training sequence of LSP as input, assigning to each input LSP vector a partition corresponding to the obtained TCQ path, taking a proper mean of each partition value assigning said mean to the corresponding path state branch.
- Method according to claim 1 characterized by the fact that a TCVQ tecnique is used instead of a TCQ one.
- Speech coder based on LPC tecniques comprising means for quantizing said LPC information characterized by the use of TCQ tecnique.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
ITMI930406A IT1271959B (en) | 1993-03-03 | 1993-03-03 | LINEAR PREDICTION SPEAKING CODEC EXCITED BY A BOOK OF CODES |
ITMI930406 | 1993-03-03 |
Publications (3)
Publication Number | Publication Date |
---|---|
EP0614075A2 true EP0614075A2 (en) | 1994-09-07 |
EP0614075A3 EP0614075A3 (en) | 1995-08-02 |
EP0614075B1 EP0614075B1 (en) | 2000-06-21 |
Family
ID=11365221
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP94103204A Expired - Lifetime EP0614075B1 (en) | 1993-03-03 | 1994-03-03 | Method and apparatus for speech coding using Trellis Coded Quantization for Linear Predictive Coding quantization |
Country Status (3)
Country | Link |
---|---|
EP (1) | EP0614075B1 (en) |
DE (1) | DE69424960T2 (en) |
IT (1) | IT1271959B (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0708435A1 (en) * | 1994-10-18 | 1996-04-24 | Matsushita Electric Industrial Co., Ltd. | Encoding and decoding apparatus of line spectrum pair parameters |
EP1072103A1 (en) * | 1998-04-14 | 2001-01-31 | Motorola, Inc. | Method and apparatus for quantizing a signal in a digital system |
CN110853659A (en) * | 2014-03-28 | 2020-02-28 | 三星电子株式会社 | Quantization apparatus for encoding an audio signal |
US11922960B2 (en) | 2014-05-07 | 2024-03-05 | Samsung Electronics Co., Ltd. | Method and device for quantizing linear predictive coefficient, and method and device for dequantizing same |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0230001A1 (en) * | 1985-12-17 | 1987-07-29 | CSELT Centro Studi e Laboratori Telecomunicazioni S.p.A. | Method of and device for speech signal coding and decoding by subband analysis and vector quantization with dynamic bit allocation |
US4975956A (en) * | 1989-07-26 | 1990-12-04 | Itt Corporation | Low-bit-rate speech coder using LPC data reduction processing |
-
1993
- 1993-03-03 IT ITMI930406A patent/IT1271959B/en active IP Right Grant
-
1994
- 1994-03-03 EP EP94103204A patent/EP0614075B1/en not_active Expired - Lifetime
- 1994-03-03 DE DE69424960T patent/DE69424960T2/en not_active Expired - Fee Related
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0230001A1 (en) * | 1985-12-17 | 1987-07-29 | CSELT Centro Studi e Laboratori Telecomunicazioni S.p.A. | Method of and device for speech signal coding and decoding by subband analysis and vector quantization with dynamic bit allocation |
US4975956A (en) * | 1989-07-26 | 1990-12-04 | Itt Corporation | Low-bit-rate speech coder using LPC data reduction processing |
Non-Patent Citations (4)
Title |
---|
ADVANCES IN SPEECH CODING, 1 January 1991 pages 47 - 56 M.W. MARCELLIN ET AL. 'A Trellis-searched 16 kbit/sec speech coder with low-delay' * |
ICASSP 91, vol.1, 14 May 1991, TORONTO pages 661 - 664 K.K. PALIWAL ET AL. 'Efficient vector quantization of LPC parameters at 24 bits/frame' * |
IEEE TRANSACTIONS ON COMMUNICATIONS, vol.38, no.1, January 1990 pages 82 - 93 M.W.MARCELLIN ET AL. 'Trellis Coded Quantization of Memoryless and Gauss-Markov sources' * |
K. SAM SHANMUGAM: "Digital and Analog Communication Systems" , , JOHN WILEY & SONS, NEW YORK * |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP0708435A1 (en) * | 1994-10-18 | 1996-04-24 | Matsushita Electric Industrial Co., Ltd. | Encoding and decoding apparatus of line spectrum pair parameters |
US5802487A (en) * | 1994-10-18 | 1998-09-01 | Matsushita Electric Industrial Co., Ltd. | Encoding and decoding apparatus of LSP (line spectrum pair) parameters |
USRE40968E1 (en) * | 1994-10-18 | 2009-11-10 | Panasonic Corporation | Encoding and decoding apparatus of LSP (line spectrum pair) parameters |
EP1072103A1 (en) * | 1998-04-14 | 2001-01-31 | Motorola, Inc. | Method and apparatus for quantizing a signal in a digital system |
EP1072103A4 (en) * | 1998-04-14 | 2008-02-06 | Motorola Inc | Method and apparatus for quantizing a signal in a digital system |
CN110853659A (en) * | 2014-03-28 | 2020-02-28 | 三星电子株式会社 | Quantization apparatus for encoding an audio signal |
US11848020B2 (en) | 2014-03-28 | 2023-12-19 | Samsung Electronics Co., Ltd. | Method and device for quantization of linear prediction coefficient and method and device for inverse quantization |
CN110853659B (en) * | 2014-03-28 | 2024-01-05 | 三星电子株式会社 | Quantization apparatus for encoding an audio signal |
US11922960B2 (en) | 2014-05-07 | 2024-03-05 | Samsung Electronics Co., Ltd. | Method and device for quantizing linear predictive coefficient, and method and device for dequantizing same |
Also Published As
Publication number | Publication date |
---|---|
EP0614075A3 (en) | 1995-08-02 |
ITMI930406A1 (en) | 1994-09-03 |
DE69424960T2 (en) | 2001-01-11 |
DE69424960D1 (en) | 2000-07-27 |
IT1271959B (en) | 1997-06-10 |
EP0614075B1 (en) | 2000-06-21 |
ITMI930406A0 (en) | 1993-03-03 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US5271089A (en) | Speech parameter encoding method capable of transmitting a spectrum parameter at a reduced number of bits | |
US6751587B2 (en) | Efficient excitation quantization in noise feedback coding with general noise shaping | |
KR100492965B1 (en) | Fast search method for nearest neighbor vector quantizer | |
US8712764B2 (en) | Device and method for quantizing and inverse quantizing LPC filters in a super-frame | |
US6161086A (en) | Low-complexity speech coding with backward and inverse filtered target matching and a tree structured mutitap adaptive codebook search | |
CA2037475C (en) | Method for reducing the search complexity in analysis-by-synthesis coding | |
US20070233473A1 (en) | Multi-path trellis coded quantization method and multi-path coded quantizer using the same | |
EP0614075B1 (en) | Method and apparatus for speech coding using Trellis Coded Quantization for Linear Predictive Coding quantization | |
US7206740B2 (en) | Efficient excitation quantization in noise feedback coding with general noise shaping | |
KR100465316B1 (en) | Speech encoder and speech encoding method thereof | |
Bouzid et al. | Optimized trellis coded vector quantization of LSF parameters, application to the 4.8 kbps FS1016 speech coder | |
US7110942B2 (en) | Efficient excitation quantization in a noise feedback coding system using correlation techniques | |
Cao et al. | A fast search algorithm for vector quantization using a directed graph | |
Aggarwal et al. | Near-optimal selection of encoding parameters for audio coding | |
Lahouti et al. | Quantization of LSF parameters using a trellis modeling | |
EP0483882B1 (en) | Speech parameter encoding method capable of transmitting a spectrum parameter with a reduced number of bits | |
US8560306B2 (en) | Method and apparatus to search fixed codebook using tracks of a trellis structure with each track being a union of tracks of an algebraic codebook | |
Mohammadi et al. | Low cost vector quantization methods for spectral coding in low rate speech coders | |
Popescu et al. | CELP coding using trellis-coded vector quantization of the excitation | |
EP0755047B1 (en) | Speech parameter encoding method capable of transmitting a spectrum parameter at a reduced number of bits | |
Ghido et al. | Optimization-quantization for least squares estimates and its application for lossless audio compression | |
Chatterjee et al. | A mixed-split scheme for 2-D DPCM based LSF quantization | |
CA2494946C (en) | Speech coder and speech decoder | |
Nurminen | Multi-mode quantization of adjacent speech parameters using a low-complexity prediction scheme. | |
Mohammadi | Combined scalar-vector quantization: a new spectral coding method for low rate speech coding |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
AK | Designated contracting states |
Kind code of ref document: A2 Designated state(s): DE FR GB IT |
|
PUAL | Search report despatched |
Free format text: ORIGINAL CODE: 0009013 |
|
AK | Designated contracting states |
Kind code of ref document: A3 Designated state(s): DE FR GB IT |
|
17P | Request for examination filed |
Effective date: 19960111 |
|
RHK1 | Main classification (correction) |
Ipc: G10L 9/14 |
|
16A | New documents despatched to applicant after publication of the search report | ||
17Q | First examination report despatched |
Effective date: 19980803 |
|
GRAG | Despatch of communication of intention to grant |
Free format text: ORIGINAL CODE: EPIDOS AGRA |
|
GRAG | Despatch of communication of intention to grant |
Free format text: ORIGINAL CODE: EPIDOS AGRA |
|
RTI1 | Title (correction) |
Free format text: METHOD AND APPARATUS FOR SPEECH CODING USING TRELLIS CODED QUANTIZATION FOR LINEAR PREDICTIVE CODING QUANTIZATION |
|
GRAG | Despatch of communication of intention to grant |
Free format text: ORIGINAL CODE: EPIDOS AGRA |
|
GRAH | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOS IGRA |
|
RIC1 | Information provided on ipc code assigned before grant |
Free format text: 7G 10L 19/04 A |
|
GRAH | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOS IGRA |
|
RAP1 | Party data changed (applicant data changed or rights of an application transferred) |
Owner name: ALCATEL |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): DE FR GB IT |
|
ET | Fr: translation filed | ||
ITF | It: translation for a ep patent filed |
Owner name: BORSANO CORRADO |
|
REF | Corresponds to: |
Ref document number: 69424960 Country of ref document: DE Date of ref document: 20000727 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: GB Payment date: 20010214 Year of fee payment: 8 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20010313 Year of fee payment: 8 Ref country code: DE Payment date: 20010313 Year of fee payment: 8 |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
26N | No opposition filed | ||
REG | Reference to a national code |
Ref country code: GB Ref legal event code: IF02 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: GB Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20020303 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: DE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20021001 |
|
GBPC | Gb: european patent ceased through non-payment of renewal fee |
Effective date: 20020303 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: FR Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20021129 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: ST |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES;WARNING: LAPSES OF ITALIAN PATENTS WITH EFFECTIVE DATE BEFORE 2007 MAY HAVE OCCURRED AT ANY TIME BEFORE 2007. THE CORRECT EFFECTIVE DATE MAY BE DIFFERENT FROM THE ONE RECORDED. Effective date: 20050303 |