WO2006075563A1 - Dispositif de codage audio, methode de codage audio et programme de codage audio - Google Patents

Dispositif de codage audio, methode de codage audio et programme de codage audio Download PDF

Info

Publication number
WO2006075563A1
WO2006075563A1 PCT/JP2006/300112 JP2006300112W WO2006075563A1 WO 2006075563 A1 WO2006075563 A1 WO 2006075563A1 JP 2006300112 W JP2006300112 W JP 2006300112W WO 2006075563 A1 WO2006075563 A1 WO 2006075563A1
Authority
WO
WIPO (PCT)
Prior art keywords
high frequency
noise level
frequency component
signal
correction coefficient
Prior art date
Application number
PCT/JP2006/300112
Other languages
English (en)
Japanese (ja)
Inventor
Osamu Shimada
Original Assignee
Nec Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nec Corporation filed Critical Nec Corporation
Priority to EP06702057.8A priority Critical patent/EP1840874B1/fr
Priority to JP2006552903A priority patent/JP5224017B2/ja
Publication of WO2006075563A1 publication Critical patent/WO2006075563A1/fr

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/04Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using predictive techniques
    • G10L19/16Vocoder architecture
    • G10L19/18Vocoders using multiple modes
    • G10L19/20Vocoders using multiple modes using sound class specific coding, hybrid encoders or object based coding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L19/00Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis
    • G10L19/02Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders
    • G10L19/0204Speech or audio signals analysis-synthesis techniques for redundancy reduction, e.g. in vocoders; Coding or decoding of speech or audio signals, using source filter models or psychoacoustic analysis using spectral analysis, e.g. transform vocoders or subband vocoders using subband decomposition
    • G10L19/0208Subband vocoders
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/038Speech enhancement, e.g. noise reduction or echo cancellation using band spreading techniques

Definitions

  • Audio encoding apparatus Audio encoding apparatus, audio encoding method, and audio encoding program
  • the present invention relates to an audio encoding device, an audio encoding method, and an audio encoding program, and in particular, an audio encoding device and an audio encoding method for encoding a wideband audio signal with a small amount of information and high quality. And audio encoding program.
  • SBR is intended to compensate for high frequency band signals (high frequency components) that are lost by audio code processing such as AAC, or band-limiting processing based on it, and is compensated by SBR.
  • the information encoded by the SBR includes information for generating a pseudo high-frequency component based on the low-frequency component transmitted using other means. By adding a pseudo high-frequency component to, the sound quality degradation due to band limitation is compensated.
  • FIG. 6 is a diagram illustrating an example of a band extension coding / decoding device using SBR.
  • the encoding side is composed of an input signal dividing unit 100, a low frequency component encoding unit 101, a high frequency component encoding unit 102, and a bit stream multiplexing unit 103
  • the decoding side is a bit stream separating unit 200, a low frequency component A component decoding unit 201, a subband division unit 202, a band extension unit 203, and a subband synthesis unit 204 are configured.
  • the input signal dividing unit 100 analyzes the input signal 1000 and outputs a high frequency sub-band signal 1001 divided into a plurality of high frequency bands and a low frequency signal 1002 including a low frequency component. .
  • the low-frequency signal 1002 is encoded into the low-frequency component information 1004 by the low-frequency component encoding unit 101 using the above-described encoding method such as AAC, and transmitted to the bit stream multiplexing unit 103.
  • the high frequency component encoding unit 102 extracts the high frequency energy information 1102 and the additional signal information 1103 from the high frequency subband signal 1001 and transmits them to the bit stream multiplexing unit 103.
  • the bitstream multiplexing unit 103 multiplexes the high frequency component information composed of the low frequency component information 1004, the high frequency energy information 1102, and the additional signal information 1103, and outputs the multiplexed bit stream 1005.
  • the high frequency energy information 1102 and the additional signal information 1103 are calculated in units of frames for each subband, for example. In consideration of the characteristics of the input signal 1000 in the time and frequency directions, it may be calculated in units of time obtained by further subdividing the frame in the time direction and in units of bands in which a plurality of subbands are combined in the frequency direction.
  • the high-frequency energy information 1102 and the additional signal information 1103 are calculated in time units obtained by further subdividing the frame in the time direction, the time change of the high-frequency subband signal 1001 can be expressed in more detail.
  • the total bits required to code the high frequency energy information 1102 and the additional signal information 1103 The number can be reduced.
  • the division unit in the time and frequency direction used for calculating the high-frequency energy information 1102 and the additional signal information 1103 is called a time-frequency grid, and the information is included in the high-frequency energy information 1102 and the additional signal information 1103.
  • the information included in the high frequency energy information 1102 and the additional signal information 1103 is only the high frequency energy information and the additional signal information. For this reason, the amount of information (total number of bits) is small compared to low band component information including waveform information and spectrum information of narrowband signals. I only need it. Therefore, it is suitable for low bit rate codes for wideband signals.
  • the multiplexed bit stream 1005 is separated into low-frequency component information 1007, high-frequency energy information 1105, and additional signal information 1106 by the bit stream separation unit 200.
  • the low-frequency component information 1007 is information encoded using an encoding method such as AAC, for example, and is decoded by the low-frequency component decoding unit 201 to generate a narrowband signal 1008 representing the low-frequency component.
  • Narrowband signal 1008 is divided into lowband subband signal 1009 by subband division section 202 and input to band extension section 203.
  • the low-frequency subband signal 1009 is also supplied to the subband synthesis unit 204 at the same time.
  • the band extension unit 203 reproduces the high frequency component lost due to the band limitation by copying the low frequency subband signal 1009 to the high frequency subband.
  • the high band energy information 1105 input to the band extension unit 203 includes energy information of the high band subband to be reproduced. After adjusting the energy of the low frequency subband signal 1009 using the high frequency energy information 1105, it is used as the high frequency component.
  • Band extension section 203 generates an additional signal in accordance with additional signal information included in additional signal information 1106. Here, a sine wave tone signal or a noise signal is used as the generated additional signal. The additional signal is added to the energy-adjusted high frequency component to obtain a high frequency sub-band signal 101.
  • Subband synthesizing section 204 band-synthesizes low band subband signal 1009 supplied from subband dividing section 202 and high band subband signal 1010 supplied from band extending section 203 to generate output signal 1011.
  • the gains of the copied low frequency subband signal 1009 and the additional signal are adjusted so that the energy of the high frequency subband signal 1010 becomes the energy value represented by the high frequency energy information 1105 (hereinafter referred to as target energy).
  • the high frequency subband signal 1010 is generated by adding the force to the high frequency component after energy adjustment.
  • the gain of the copied low-frequency subband signal 1009 and the additional signal can be determined, for example, by the following procedure.
  • one of the copied low-frequency subband signal 1009 and the additional signal is set as a main component of the high-frequency subband signal 1010, and the other as a subcomponent.
  • G is the main component amplitude adjustment gain
  • G is the sub component amplitude adjustment gain
  • E and N are the main component amplitude adjustment gain
  • Represent the energies of the low frequency subband signal 1009 and the additional signal, respectively. If the energy of the additional signal is normally set to 1, N l. R represents the target energy of the high frequency sub-band signal 1010, Q represents the energy ratio of the main component and the sub component, and R and Q are included in the high frequency energy information 1105 and the additional signal information 1106. Note that sqrt (') is an operator for finding the square root. On the other hand, when the additional signal is the main component and the low-frequency subband signal 1009 is the main component, the gain is determined by the following equation.
  • the low frequency subband signal 1009 and the additional signal are weighted and added to calculate the high frequency subband signal 1010.
  • an accurate energy ratio Q between the low-frequency subband signal 1009 and the noise signal to be added is added to the additional signal information 1103 generated on the code side.
  • the high frequency component code key unit 102 needs to accurately calculate the noise level of the high frequency component in the input signal.
  • Non-Patent Document 3 discloses a first conventional example of a high-frequency component code key unit 102 that calculates a noise level of a high-frequency component. 7 includes a time Z frequency dial generation unit 300, a spectrum envelope calculation unit 301, a noise level calculation unit 302, and a noise level integration unit 303.
  • the time Z frequency grid generation unit 300 uses the high frequency subband signal 1001 to Group multiple subband signals in the frequency direction and time Z frequency grid information no
  • the spectrum envelope calculation unit 301 extracts the target energy R of the high-frequency subband signal in units of time Z frequency grids, and supplies the target energy R to the bitstream multiplexing unit 103 as high-frequency energy information 1102.
  • the noise level calculation unit 302 outputs the ratio of the noise component included in the subband signal as the noise level 1101 for each subband.
  • the noise level integration unit 303 uses the average value of the noise levels in a plurality of subbands to obtain additional signal information 1103 representing the energy ratio Q in units of time Z frequency grids, and supplies the additional signal information 1103 to the bit stream multiplexing unit 103. To do.
  • X (k, 1) and Y (k, 1) denote the subband signal and the predicted subband signal of subband k, respectively.
  • a linear prediction method using a covariance method or an autocorrelation method is known.
  • the difference between the subband signal X and the predicted subband signal Y becomes small, and the value of the noise level T (k) becomes large.
  • the difference between the predicted subband signal Y and the subband signal X increases, and the value of the noise level T (k) decreases. In this way, the noise level T (k) can be calculated based on the size of the noise component contained in the subband signal.
  • the noise level integration unit 303 calculates the energy ratio Q between the low frequency subband signal and the noise signal in units of a plurality of subbands based on the time Z frequency grid information 1100. This is more than one subband unit, rather than calculating and signing the energy ratio Q for each subband unit. This is because the number of bits required for the additional signal information 1103 can be reduced by calculating the energy ratio Q. For example, N subbands from subband k to subband k + N— 1
  • the additional signal information 1103 is obtained by averaging the noise levels 1101 of N subbands from subband k to subband k + N ⁇ 1.
  • fNoise the frequency number of the additional signal information 1103, and c is a constant.
  • Patent Document 1 Japanese Translation of Special Publication 2002-536679
  • Non-Patent Document 1 "Digital Radio Musice (DRM); System Specification", E TSI, TS 101 980 VI. 1. 1, 5. 2. 6, September 2001
  • Non-Patent Document 2 AES (Audio Engineering Society) Convention Paper 55 53 ", 112th AES Convention, May 2002
  • Non-Patent Document 3 "Enhanced aacPlus general audio codec; Enhanced aacPl us encoder SBR part", 3GPP, TS 26. 404 V6. 0. 0, September 2004 Disclosure of Invention
  • the conventional additional signal information calculation method averages the noise level calculated independently for each subband, the perceptual priority of the subband is not considered. Therefore, the noise level of the sub-bands that are important to the auditory sense depends on the importance of the additional signal information.
  • the high-quality audio signal encoding device cannot be realized.
  • the method of calculating the additional signal information using the spectrum envelope has a problem that the amount of calculation increases because high resolution frequency analysis and smoothing processing are required.
  • the noise level varies greatly depending on the level of smoothness, and it is difficult to optimize the level of smoothing.
  • the present invention was invented in view of the above problems, and its purpose is to reduce the amount of additional signal information that reflects the noise level of an audibly important subband according to the degree of importance.
  • the object is to provide a technique relating to a high-quality audio signal code that can be calculated in quantity.
  • a first invention for solving the above-described problem is an input signal dividing unit for extracting a high frequency signal from an input signal, and generating a first high frequency component information by extracting a spectrum of the high frequency signal.
  • a first high-frequency component encoding unit ; a noise level calculation unit that obtains the noise level of the high-frequency signal by reflecting the importance of each frequency component; and second high-frequency component information using the noise level.
  • a second invention that solves the above-described problem is an input signal dividing unit that extracts a high-frequency signal from an input signal, and generates a first high-frequency component information by extracting a spectrum of the high-frequency signal.
  • a first high-frequency component encoding unit; a noise level calculation unit that calculates a noise level using the high-frequency signal; a correction coefficient calculation unit that calculates a correction coefficient using the high-frequency signal; and the correction A noise level correction unit that corrects the noise level using a coefficient to obtain a corrected noise level; and a second high frequency component encoding unit that generates second high frequency component information using the corrected noise level;
  • An audio encoding device comprising: a bit stream multiplexing unit that multiplexes the first high frequency component information and the second high frequency component information and outputs a multiplexed bit stream. It is.
  • a third invention for solving the above-described problem is the correction coefficient calculation according to the second invention.
  • the unit calculates a correction coefficient that reflects the importance of each frequency component of the high frequency signal.
  • the correction coefficient calculation unit calculates energy for each frequency band of the high-frequency signal, and based on the energy for each frequency band. A correction coefficient is calculated.
  • a fifth invention for solving the above-mentioned problems is characterized in that, in the second or third invention, the correction coefficient calculation unit calculates a correction coefficient having a small value at a high frequency. To do.
  • a sixth invention that solves the above-described problem is that, in the first invention, at least the noise level calculated by the noise level calculation unit reflecting the importance of each frequency component of the high-frequency signal. It is characterized by smoothing in the time direction or frequency direction.
  • the correction coefficient calculation unit calculates the frequency component according to each frequency component of the high-frequency signal.
  • the correction coefficient is smoothed at least in the time direction or the frequency direction.
  • An eighth invention for solving the above-mentioned problem is to extract a high frequency signal from an input signal, extract a spectrum of the high frequency signal to generate first high frequency component information, and The noise level is calculated by reflecting the importance of each frequency component, second high frequency component information is generated from the noise level, and the first high frequency component information and the second high frequency component information are generated. And an audio coding method characterized by outputting a multiplexed bit stream. It is.
  • a ninth invention for solving the above-described problem is to extract a high frequency signal from an input signal, extract a spectrum of the high frequency signal to generate first high frequency component information, and The noise level is obtained using the high frequency signal, the correction coefficient is obtained using the high frequency signal, the noise level is corrected using the correction coefficient to obtain the correction noise level, and the second high noise is obtained using the correction noise level.
  • An audio coding method characterized by generating band component information, multiplexing the first high band component information and the second high band component information, and outputting a multiplexed bit stream.
  • a tenth invention for solving the above-mentioned problems is the above-mentioned eighth invention, wherein the correction coefficient is When obtaining the correction coefficient, a correction coefficient is obtained corresponding to the auditory importance corresponding to each frequency component of the high frequency signal.
  • a twelfth invention for solving the above-mentioned problems is characterized in that, in the above-mentioned eighth or ninth invention, when obtaining the correction coefficient, a correction coefficient having a small value at a high frequency is calculated. To do.
  • the eighth invention at the time of obtaining the noise level, at least a noise level obtained by reflecting importance of each frequency component of the high frequency signal is used. It is characterized by smoothing in the time direction or frequency direction.
  • the correction coefficient when the correction coefficient is obtained, the correction coefficient is determined according to each frequency component of the high frequency signal.
  • the calculated correction coefficient is smoothed at least in the time direction or frequency direction.
  • a fifteenth aspect of the present invention for solving the above-described problem is a process of extracting a high frequency signal from an input signal, a process of extracting a spectrum of the high frequency signal and generating first high frequency component information, Processing for obtaining the noise level of the high frequency signal by reflecting the importance of each frequency component, processing for generating second high frequency component information using the noise level, and the first high frequency component information
  • the present invention is configured to calculate a correction coefficient corresponding to auditory importance using a high frequency sub-band signal, correct a noise level, and generate additional signal information. It is possible to accurately reflect the noise level of a sub-band that is important perceptually. For this reason, a high-quality audio encoding device can be realized.
  • the invention's effect it is possible to calculate a correction coefficient based on the perceptual importance of the input signal and correct the noise level of each subband.
  • the correction coefficient calculation of the present invention performs frequency analysis with normal resolution, the amount of computation required for high-resolution frequency analysis is reduced, and subband noise that reflects auditory importance is reflected. You can ask for the level. As a result, a high-quality audio encoding device can be realized.
  • FIG. 1 is a block diagram showing the configuration of the best mode for carrying out the first invention of the present invention.
  • FIG. 2 is an explanatory diagram showing an operation concept of a correction coefficient calculation unit in the present invention.
  • FIG. 3 is a block diagram showing a configuration of an input signal dividing unit.
  • FIG. 4 is a block diagram showing the configuration of the best mode for carrying out the second invention of the present invention.
  • FIG. 5 is a block diagram showing the configuration of the best mode for carrying out the third invention of the present invention.
  • FIG. 6 is a block diagram showing a band extension code decoding apparatus.
  • FIG. 7 is a block diagram showing a configuration of a high frequency component code key unit.
  • an audio encoding device includes an input signal dividing unit 100, a low frequency component encoding unit 101, a time Z frequency grid generating unit 300, a spectrum envelope.
  • the high frequency component code key unit 102 and the high frequency component code key unit 500 are different. Comparing these components in more detail using FIG. 1 and FIG. 7, a correction coefficient calculation unit 400 and a noise level correction unit 401 are added to the high frequency component code unit 500, and the noise level integration unit 300 is replaced by the noise level integration unit 402.
  • the correction coefficient calculation unit 400, the noise level correction unit 401, and the noise level integration unit 402 will be described.
  • the time Z frequency grid information 1100 obtained by grouping a plurality of subband signals in the time and frequency directions in the time Z frequency grid generation unit 300 using the high frequency subband signal 1001 This is transmitted to the correction coefficient calculation unit 400.
  • the correction coefficient calculation unit 400 calculates the perceptual importance of each subband using the high frequency subband signal 1001 and the time Z frequency grid information 1100, and calculates the correction coefficient 1200 of each subband to the noise level correction unit 401. Communicate to.
  • the noise level 1101 of each subband calculated by the noise level calculation unit 302 using the high frequency subband signal 1001 is also transmitted.
  • the noise level correction unit 401 corrects the noise level 1101 of each subband based on the correction coefficient 1200, and outputs the corrected noise level 1201 to the noise level integration unit 402.
  • Noise level integration section 402 calculates an average value of corrected noise levels 1103 in a plurality of subbands based on time Z frequency grid information 1100.
  • the energy ratio of the noise component is calculated in units of time Z frequency grid and output as additional signal information 1103.
  • FIG. 2 shows a part of the spectrum when the input signal 1000 is subjected to frequency analysis, where the horizontal axis represents frequency and the vertical axis represents energy.
  • N subbands from subband k to subband k + N— 1 are paired.
  • the noise component energy ratio Q in region 2 must be reflected in the additional signal information 1103 in accordance with the importance of region 2. In order to do so, it is necessary to calculate the listening and emotional importance of each subband.
  • the correction coefficient 1200 representing the perceptual importance of each subband can be calculated according to the energy of the high frequency subband signal 1001, for example.
  • Subband k to subband k
  • the correction coefficient a (k) for subband k can be expressed by the following equation, for example: .
  • E represents the energy of each subband.
  • the energy of each subband may be calculated for each time grid included in the time Z frequency grid information 1100, or may be calculated using subband signals included in a plurality of time grids.
  • the energy of the high frequency sub-band signal 1001 is used as it is, but the energy of the sub-band signal 1101 may be corrected.
  • the energy of the sub-band signal can be used for calculating the force correction coefficient in a logarithm rather than using it as it is.
  • the correction coefficient may be calculated by positively using auditory characteristics. For example, simultaneous masking that makes it impossible to perceive small sounds that are present simultaneously with loud sounds, It is also possible to calculate a correction coefficient that takes into account the effect of the continuous masking that occurs. Sounds smaller than the masking threshold cannot be perceived. Therefore, the correction coefficient corresponding to the auditory importance can be calculated by relatively reducing the correction coefficient of the subband that can be ignored for auditory perception. Conversely, the correction coefficient for subbands larger than the masking threshold may be relatively large.
  • the noise level correction unit 401 corrects the noise level 1101 of each subband calculated by the noise level calculation unit based on the correction coefficient 1200 calculated by the correction coefficient calculation unit, and the corrected noise level 1201 is input to the noise level integration unit 303. Output.
  • the corrected noise level T (k) is given by
  • T (k) a (k) X T (k)
  • the result obtained by adding a constant to the product can also be used as a corrected noise level. Furthermore, by defining the correction noise level as an arbitrary function of the correction factor 1200 and the noise level 1101.
  • the noise level integration unit 402 uses the corrected noise level 1201 to calculate the energy ratio Q of the additional signal for each frequency grid included in the time Z frequency grid information 1100 and outputs it as the attached calorie signal information 1103. For example, from subband k to subband k + N— 1
  • the input signal dividing unit 100 can be configured with a subband dividing unit 110 and a subband combining unit 111 as shown in FIG.
  • Subband dividing section 110 divides input signal 1000 into N subbands and outputs high frequency subband signal 1001.
  • the subband synthesizing unit 111 generates a low frequency signal 1002 by performing subband synthesis using M (MMN) subband signals of the low frequency of the subband signal.
  • M M
  • the down-sampling filter 112 can be used to down-sample the input signal 1000.
  • the down-sampling filter 112 includes a low-pass filter having a pass band corresponding to the band of the low-frequency signal 1002, and performs high-frequency suppression processing using the low-frequency filter before down-sampling processing. Further, as shown in FIG. 3 (c), the input signal 1000 may be output as the low frequency signal 1002 without being processed.
  • correction coefficient 1200 corresponding to auditory importance is calculated using high frequency subband signal 1001, noise level 1101 is corrected, and additional signal information 1103 is generated. Therefore, it is possible to accurately reflect the noise level of the sub-band that is important perceptually. As a result, a high-quality audio encoding device can be realized.
  • the best mode for carrying out the second invention of the present invention is that an input signal dividing unit 100, a low frequency component encoding unit 101, a time Z frequency grid generating unit 300, a spectral packet, An envelope calculation unit 301, a noise level calculation unit 302, a correction coefficient calculation unit 403, a noise level correction unit 401, a noise level integration unit 402, and a bit stream multiplexing unit 103 are included.
  • the correction coefficient calculation unit 400 is simply replaced with the correction coefficient calculation unit 403, and the other Part Are exactly the same. Therefore, the correction coefficient calculation unit 403 will be described in detail.
  • the correction coefficient calculation unit 403 calculates a correction coefficient 1202 by a predetermined method based on the time Z frequency grid information 1100 and outputs the correction coefficient 1202 to the noise level correction unit 401.
  • the correction coefficient 1202 can be calculated such that the correction coefficient 1202 takes a small value for a high frequency.
  • the correspondence relationship between the frequency and the correction coefficient 1202 can be determined to be expressed by a linear function as the simplest example, or may be determined to be expressed by a nonlinear function.
  • a high-frequency signal component is often attenuated more than a low-frequency signal component. Therefore, high-quality additional signal information 1103 can be calculated using the above-described method.
  • this embodiment uses a correction coefficient 1202 based on the characteristics of a general audio signal, the amount of calculation can be reduced as compared with the first embodiment of the present invention.
  • the third embodiment of the present invention is a computer that operates according to program 601 when the above-described first and second embodiments of the present invention are configured with program 601.
  • FIG. 5 the third embodiment of the present invention is a computer that operates according to program 601 when the above-described first and second embodiments of the present invention are configured with program 601.
  • the program 601 is read into a computer 600 (central processing unit; processor; data processing device) and controls the operation of the computer 600 (central processing unit; processor; data processing device).
  • the computer 600 central processing unit; processor; data processing unit

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Spectroscopy & Molecular Physics (AREA)
  • Quality & Reliability (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

Au moyen d'un signal de sous-bande à grande portée, un coefficient de correction correspondant à l'intensité du sens auditif est calculé pour corriger un niveau de bruit et générer une information de signal additionnelle, reflétant ainsi, de manière précise, le niveau de bruit de la sous-bande intense dans le sens auditif. Ainsi, il est possible de calculer l'information de signal additionnelle reflétant le niveau de bruit de la sous-bande intense dans le sens auditif selon l'intensité par une petite quantité de calculs. La quantité de calculs peut être encore réduite au moyen d'un coefficient de correction basé sur la caractéristique d'un signal audio ordinaire.
PCT/JP2006/300112 2005-01-11 2006-01-06 Dispositif de codage audio, methode de codage audio et programme de codage audio WO2006075563A1 (fr)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP06702057.8A EP1840874B1 (fr) 2005-01-11 2006-01-06 Dispositif de codage audio, methode de codage audio et programme de codage audio
JP2006552903A JP5224017B2 (ja) 2005-01-11 2006-01-06 オーディオ符号化装置、オーディオ符号化方法およびオーディオ符号化プログラム

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
JP2005-003291 2005-01-11
JP2005003291 2005-01-11

Publications (1)

Publication Number Publication Date
WO2006075563A1 true WO2006075563A1 (fr) 2006-07-20

Family

ID=36677588

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/JP2006/300112 WO2006075563A1 (fr) 2005-01-11 2006-01-06 Dispositif de codage audio, methode de codage audio et programme de codage audio

Country Status (4)

Country Link
US (1) US8082156B2 (fr)
EP (1) EP1840874B1 (fr)
JP (1) JP5224017B2 (fr)
WO (1) WO2006075563A1 (fr)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011086924A1 (fr) * 2010-01-14 2011-07-21 パナソニック株式会社 Appareil de codage audio et procédé de codage audio
WO2012050023A1 (fr) * 2010-10-15 2012-04-19 ソニー株式会社 Dispositif et procédé de codage, dispositif et procédé de décodage, et programme
JP2015228044A (ja) * 2007-05-08 2015-12-17 サムスン エレクトロニクス カンパニー リミテッド オーディオ信号の符号化及び復号化方法並びにその装置
JP2016006540A (ja) * 2009-10-07 2016-01-14 ソニー株式会社 復号装置および方法、並びにプログラム
US9583112B2 (en) 2010-04-13 2017-02-28 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US9679580B2 (en) 2010-04-13 2017-06-13 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US9691410B2 (en) 2009-10-07 2017-06-27 Sony Corporation Frequency band extending device and method, encoding device and method, decoding device and method, and program
US9875746B2 (en) 2013-09-19 2018-01-23 Sony Corporation Encoding device and method, decoding device and method, and program
US10692511B2 (en) 2013-12-27 2020-06-23 Sony Corporation Decoding apparatus and method, and program

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2005111568A1 (fr) * 2004-05-14 2005-11-24 Matsushita Electric Industrial Co., Ltd. Dispositif de codage, dispositif de décodage et méthode pour ceux-ci
JPWO2008053970A1 (ja) * 2006-11-02 2010-02-25 パナソニック株式会社 音声符号化装置、音声復号化装置、およびこれらの方法
US7853545B2 (en) * 2007-02-26 2010-12-14 International Business Machines Corporation Preserving privacy of one-dimensional data streams using dynamic correlations
US7840516B2 (en) * 2007-02-26 2010-11-23 International Business Machines Corporation Preserving privacy of one-dimensional data streams by perturbing data with noise and using dynamic autocorrelation
JP2009044268A (ja) * 2007-08-06 2009-02-26 Sharp Corp 音声信号処理装置、音声信号処理方法、音声信号処理プログラム、及び、記録媒体
US8190440B2 (en) * 2008-02-29 2012-05-29 Broadcom Corporation Sub-band codec with native voice activity detection
US20120278087A1 (en) * 2009-10-07 2012-11-01 Nec Corporation Multiband compressor and method of adjusting the same
JP5609737B2 (ja) 2010-04-13 2014-10-22 ソニー株式会社 信号処理装置および方法、符号化装置および方法、復号装置および方法、並びにプログラム
US8560330B2 (en) * 2010-07-19 2013-10-15 Futurewei Technologies, Inc. Energy envelope perceptual correction for high band coding
JP5777041B2 (ja) * 2010-07-23 2015-09-09 沖電気工業株式会社 帯域拡張装置及びプログラム、並びに、音声通信装置
JP5743137B2 (ja) * 2011-01-14 2015-07-01 ソニー株式会社 信号処理装置および方法、並びにプログラム
EP2710588B1 (fr) 2011-05-19 2015-09-09 Dolby Laboratories Licensing Corporation Détection légale de méthodes de codage audio paramétrique
JP5942358B2 (ja) 2011-08-24 2016-06-29 ソニー株式会社 符号化装置および方法、復号装置および方法、並びにプログラム
WO2013124445A2 (fr) * 2012-02-23 2013-08-29 Dolby International Ab Procédés et systèmes pour la reconstitution efficace d'un contenu audio haute fréquence
US9542955B2 (en) * 2014-03-31 2017-01-10 Qualcomm Incorporated High-band signal coding using multiple sub-bands
KR20210111603A (ko) * 2020-03-03 2021-09-13 삼성전자주식회사 음질 개선 방법 및 그 장치

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06259094A (ja) * 1993-03-05 1994-09-16 Sony Corp 圧縮データ記録及び又は再生装置又は方法
JPH06313694A (ja) * 1993-04-28 1994-11-08 Toshiba Corp 熱交換器および空調換気扇
JPH09127986A (ja) * 1995-10-26 1997-05-16 Sony Corp 符号化信号の多重化方法及び信号符号化装置
JPH09261066A (ja) * 1996-03-27 1997-10-03 Matsushita Electric Ind Co Ltd ロスレス符号装置とロスレス記録媒体とロスレス復号装置とロスレス符号復号装置
JPH11514453A (ja) * 1995-09-14 1999-12-07 エリクソン インコーポレイテッド 音声信号を適応的に濾波して雑音のある環境条件での音声了解度を高めるシステム
WO2003046891A1 (fr) * 2001-11-29 2003-06-05 Coding Technologies Ab Procede permettant d'ameliorer la reconstruction des hautes frequences

Family Cites Families (19)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2705377B2 (ja) * 1991-07-31 1998-01-28 松下電器産業株式会社 帯域分割符号化方法
JP3278900B2 (ja) * 1992-05-07 2002-04-30 ソニー株式会社 データ符号化装置及び方法
JP3186315B2 (ja) * 1993-02-27 2001-07-11 ソニー株式会社 信号圧縮装置、信号伸張装置、信号送信装置、信号受信装置及び信号送受信装置
JPH08123490A (ja) * 1994-10-24 1996-05-17 Matsushita Electric Ind Co Ltd スペクトル包絡量子化装置
JPH08162963A (ja) * 1994-11-30 1996-06-21 Sony Corp データ符号化装置および復号装置
TW321810B (fr) 1995-10-26 1997-12-01 Sony Co Ltd
EP0880235A1 (fr) 1996-02-08 1998-11-25 Matsushita Electric Industrial Co., Ltd. Codeur, decodeur, codeur-decodeur et support d'enregistrement de signal audio large bande
JP3344944B2 (ja) * 1997-05-15 2002-11-18 松下電器産業株式会社 オーディオ信号符号化装置,オーディオ信号復号化装置,オーディオ信号符号化方法,及びオーディオ信号復号化方法
EP0878790A1 (fr) * 1997-05-15 1998-11-18 Hewlett-Packard Company Système de codage de la parole et méthode
SE512719C2 (sv) 1997-06-10 2000-05-02 Lars Gustaf Liljeryd En metod och anordning för reduktion av dataflöde baserad på harmonisk bandbreddsexpansion
SE9903553D0 (sv) 1999-01-27 1999-10-01 Lars Liljeryd Enhancing percepptual performance of SBR and related coding methods by adaptive noise addition (ANA) and noise substitution limiting (NSL)
AUPR647501A0 (en) * 2001-07-19 2001-08-09 Vast Audio Pty Ltd Recording a three dimensional auditory scene and reproducing it for the individual listener
US7333929B1 (en) * 2001-09-13 2008-02-19 Chmounk Dmitri V Modular scalable compressed audio data stream
JP3926726B2 (ja) 2001-11-14 2007-06-06 松下電器産業株式会社 符号化装置および復号化装置
KR100723753B1 (ko) * 2002-08-01 2007-05-30 마츠시타 덴끼 산교 가부시키가이샤 스펙트럼 대역 복사에 의한 오디오 디코딩 장치 및 오디오디코딩 방법
JP3861770B2 (ja) * 2002-08-21 2006-12-20 ソニー株式会社 信号符号化装置及び方法、信号復号装置及び方法、並びにプログラム及び記録媒体
CA2469674C (fr) * 2002-09-19 2012-04-24 Matsushita Electric Industrial Co., Ltd. Procede et appareil de decodage audio
JP4380174B2 (ja) * 2003-02-27 2009-12-09 沖電気工業株式会社 帯域補正装置
WO2005004113A1 (fr) * 2003-06-30 2005-01-13 Fujitsu Limited Dispositif de codage audio

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH06259094A (ja) * 1993-03-05 1994-09-16 Sony Corp 圧縮データ記録及び又は再生装置又は方法
JPH06313694A (ja) * 1993-04-28 1994-11-08 Toshiba Corp 熱交換器および空調換気扇
JPH11514453A (ja) * 1995-09-14 1999-12-07 エリクソン インコーポレイテッド 音声信号を適応的に濾波して雑音のある環境条件での音声了解度を高めるシステム
JPH09127986A (ja) * 1995-10-26 1997-05-16 Sony Corp 符号化信号の多重化方法及び信号符号化装置
JPH09261066A (ja) * 1996-03-27 1997-10-03 Matsushita Electric Ind Co Ltd ロスレス符号装置とロスレス記録媒体とロスレス復号装置とロスレス符号復号装置
WO2003046891A1 (fr) * 2001-11-29 2003-06-05 Coding Technologies Ab Procede permettant d'ameliorer la reconstruction des hautes frequences

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2015228044A (ja) * 2007-05-08 2015-12-17 サムスン エレクトロニクス カンパニー リミテッド オーディオ信号の符号化及び復号化方法並びにその装置
JP2017203995A (ja) * 2007-05-08 2017-11-16 サムスン エレクトロニクス カンパニー リミテッド オーディオ信号の符号化及び復号化方法並びにその装置
US9691410B2 (en) 2009-10-07 2017-06-27 Sony Corporation Frequency band extending device and method, encoding device and method, decoding device and method, and program
JP2017203999A (ja) * 2009-10-07 2017-11-16 ソニー株式会社 復号装置および方法、並びにプログラム
JP2016006540A (ja) * 2009-10-07 2016-01-14 ソニー株式会社 復号装置および方法、並びにプログラム
JPWO2011086924A1 (ja) * 2010-01-14 2013-05-16 パナソニック株式会社 音声符号化装置および音声符号化方法
WO2011086924A1 (fr) * 2010-01-14 2011-07-21 パナソニック株式会社 Appareil de codage audio et procédé de codage audio
JP5809066B2 (ja) * 2010-01-14 2015-11-10 パナソニック インテレクチュアル プロパティ コーポレーション オブアメリカPanasonic Intellectual Property Corporation of America 音声符号化装置および音声符号化方法
US10224054B2 (en) 2010-04-13 2019-03-05 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US10297270B2 (en) 2010-04-13 2019-05-21 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US10381018B2 (en) 2010-04-13 2019-08-13 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US9583112B2 (en) 2010-04-13 2017-02-28 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US9679580B2 (en) 2010-04-13 2017-06-13 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US10546594B2 (en) 2010-04-13 2020-01-28 Sony Corporation Signal processing apparatus and signal processing method, encoder and encoding method, decoder and decoding method, and program
US9177563B2 (en) 2010-10-15 2015-11-03 Sony Corporation Encoding device and method, decoding device and method, and program
US9767824B2 (en) 2010-10-15 2017-09-19 Sony Corporation Encoding device and method, decoding device and method, and program
RU2630384C1 (ru) * 2010-10-15 2017-09-07 Сони Корпорейшн Устройство и способ декодирования и носитель записи программы
US9536542B2 (en) 2010-10-15 2017-01-03 Sony Corporation Encoding device and method, decoding device and method, and program
TWI479481B (zh) * 2010-10-15 2015-04-01 Sony Corp Encoding apparatus and method, decoding apparatus and method, and program
US10236015B2 (en) 2010-10-15 2019-03-19 Sony Corporation Encoding device and method, decoding device and method, and program
CN103155031A (zh) * 2010-10-15 2013-06-12 索尼公司 编码装置和方法、解码装置和方法以及程序
JP2012083678A (ja) * 2010-10-15 2012-04-26 Sony Corp 符号化装置および方法、復号装置および方法、並びにプログラム
WO2012050023A1 (fr) * 2010-10-15 2012-04-19 ソニー株式会社 Dispositif et procédé de codage, dispositif et procédé de décodage, et programme
US9875746B2 (en) 2013-09-19 2018-01-23 Sony Corporation Encoding device and method, decoding device and method, and program
US10692511B2 (en) 2013-12-27 2020-06-23 Sony Corporation Decoding apparatus and method, and program
US11705140B2 (en) 2013-12-27 2023-07-18 Sony Corporation Decoding apparatus and method, and program

Also Published As

Publication number Publication date
EP1840874B1 (fr) 2019-04-10
US8082156B2 (en) 2011-12-20
EP1840874A1 (fr) 2007-10-03
JP5224017B2 (ja) 2013-07-03
US20080140425A1 (en) 2008-06-12
EP1840874A4 (fr) 2008-05-28
JPWO2006075563A1 (ja) 2008-08-07

Similar Documents

Publication Publication Date Title
JP5224017B2 (ja) オーディオ符号化装置、オーディオ符号化方法およびオーディオ符号化プログラム
JP4934427B2 (ja) 音声信号復号化装置及び音声信号符号化装置
JP3579047B2 (ja) オーディオ復号装置と復号方法およびプログラム
KR101278546B1 (ko) 대역폭 확장 출력 데이터를 생성하기 위한 장치 및 방법
JP5203077B2 (ja) 音声符号化装置及び方法、音声復号化装置及び方法、並びに、音声帯域拡張装置及び方法
EP2056294B1 (fr) Appareil, support et procédé pour coder et décoder un signal haute fréquence
JP2022123060A (ja) 符号化オーディオ信号を復号する復号装置および復号方法
US10255928B2 (en) Apparatus, medium and method to encode and decode high frequency signal
KR102055022B1 (ko) 부호화 장치 및 방법, 복호 장치 및 방법, 및 프로그램
JP4313993B2 (ja) オーディオ復号化装置およびオーディオ復号化方法
JP4354561B2 (ja) オーディオ信号符号化装置及び復号化装置

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application
DPE1 Request for preliminary examination filed after expiration of 19th month from priority date (pct application filed from 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2006552903

Country of ref document: JP

WWE Wipo information: entry into national phase

Ref document number: 11794984

Country of ref document: US

WWE Wipo information: entry into national phase

Ref document number: 2006702057

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: DE

WWP Wipo information: published in national office

Ref document number: 2006702057

Country of ref document: EP