EP2807833A2 - Système de rendu audio et procédé correspondant - Google Patents

Système de rendu audio et procédé correspondant

Info

Publication number
EP2807833A2
EP2807833A2 EP13710012.9A EP13710012A EP2807833A2 EP 2807833 A2 EP2807833 A2 EP 2807833A2 EP 13710012 A EP13710012 A EP 13710012A EP 2807833 A2 EP2807833 A2 EP 2807833A2
Authority
EP
European Patent Office
Prior art keywords
audio
signal
speaker
channel
speaker arrangement
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Withdrawn
Application number
EP13710012.9A
Other languages
German (de)
English (en)
Inventor
Bernd Roland JACQUES
Aki Sakari HÄRMÄ
Mun Hum Park
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Gibson Innovations Belgium NV
Original Assignee
Koninklijke Philips NV
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Koninklijke Philips NV filed Critical Koninklijke Philips NV
Publication of EP2807833A2 publication Critical patent/EP2807833A2/fr
Withdrawn legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S5/00Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/305Electronic adaptation of stereophonic audio signals to reverberation of the listening space
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/20Arrangements for obtaining desired frequency or directional characteristics
    • H04R1/32Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
    • H04R1/34Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means
    • H04R1/345Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means for loudspeakers
    • H04R1/347Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by using a single transducer with sound reflecting, diffracting, directing or guiding means for loudspeakers for obtaining a phase-shift between the front and back acoustic wave
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R1/00Details of transducers, loudspeakers or microphones
    • H04R1/20Arrangements for obtaining desired frequency or directional characteristics
    • H04R1/32Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only
    • H04R1/40Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers
    • H04R1/403Arrangements for obtaining desired frequency or directional characteristics for obtaining desired directional characteristic only by combining a number of identical transducers loud-speakers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R2203/00Details of circuits for transducers, loudspeakers or microphones covered by H04R3/00 but not provided for in any of its subgroups
    • H04R2203/12Beamforming aspects for stereophonic sound reproduction with loudspeaker arrays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04RLOUDSPEAKERS, MICROPHONES, GRAMOPHONE PICK-UPS OR LIKE ACOUSTIC ELECTROMECHANICAL TRANSDUCERS; DEAF-AID SETS; PUBLIC ADDRESS SYSTEMS
    • H04R3/00Circuits for transducers, loudspeakers or microphones
    • H04R3/12Circuits for transducers, loudspeakers or microphones for distributing signals to two or more loudspeakers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/07Synergistic effects of band splitting and sub-band processing
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S7/00Indicating arrangements; Control arrangements, e.g. balance control
    • H04S7/30Control circuits for electronic adaptation of the sound field
    • H04S7/302Electronic adaptation of stereophonic sound system to listener position or orientation
    • H04S7/303Tracking of listener position or orientation

Definitions

  • the invention relates to an audio rendering system and in particular, but not exclusively, to a spatial audio rendering system such as a surround sound audio rendering system.
  • Multi-channel audio rendering, and in particular multi-channel spatial sound rendering, beyond simple stereo has become commonplace through applications such as surround sound home cinema systems.
  • loudspeakers positioned at specific spatial positions relative to a listening position.
  • a 5.1 home cinema system provides spatial sound via five loudspeakers being positioned with one speaker directly in front of the listening position (the center channel), one speaker to the front left of the listening position, one speaker to the front right of the listening position, one speaker to the rear left of the listening position, and one speaker to the rear right of the listening position.
  • a non-spatial low frequency speaker is often provided.
  • Such conventional systems are based on the reproduction of audio signals at specific nominal positions relative to the listening position.
  • One speaker is typically provided for each audio channel and therefore speakers must be positioned at locations corresponding to the predetermined or nominal positions for the system.
  • audio rendering systems may perform an upmixing of one or more input channels to generate additional channels.
  • the system may accordingly employ an algorithm which synthesizes additional loudspeaker driving signals from a given input audio signal.
  • an improved audio rendering approach would be advantageous and in particular an audio rendering approach that allows upmixing to synthesize one or more additional channels.
  • an audio rendering approach allowing for increased flexibility, reduced complexity, an improved user experience, a more encapsulating sound experience, reduced spatial distortions, and/or improved performance would be
  • the Invention seeks to preferably mitigate, alleviate or eliminate one or more of the above mentioned disadvantages singly or in any combination.
  • an audio rendering system comprising: an audio renderer; a first speaker arrangement coupled to the audio renderer and arranged to render audio to a listening position, the first speaker arrangement having a directional radiation pattern with a direction from the first speaker arrangement to the listening position being within a 3dB beamwidth of a main lobe of the first speaker arrangement; and a second speaker arrangement coupled to the audio renderer and arranged to render audio to the listening position, the second speaker arrangement having a directional radiation pattern with a direction from the second speaker arrangement to the listening position being outside a 3dB beamwidth of a main lobe of the second speaker arrangement; wherein the audio renderer comprises: a receiver for receiving a multi-channel audio signal; a correlation estimator for generating a correlation measure for a first channel signal and a second channel signal of the multi-channel audio signal; an upmixer for upmixing the first channel signal to a first audio signal and a second audio signal in response to a correlation measure, the second audio signal corresponding to
  • the invention may provide an improved user experience to a listener.
  • a more encapsulating and immerging user experience may often be achieved.
  • an extended sound stage can be perceived.
  • the sound stage may be perceived as natural and spatial distortions of spatially well-defined positions may be reduced.
  • the combination of an upmixing based on the correlation/coherence between two channels combined with the rendering using non-reflected and reflected paths may provide an improved perceived sound stage expansion in many implementations.
  • it may allow for a spatial expansion of ambient sound that is typically perceived as not having strong spatial cues while at the same time allowing specific and well defined individual spatial sound sources to appear unmodified.
  • the approach may specifically result in an audio rendering which expands the general ambient sound to be perceived to increasingly surround the user without changing the specific sound sources in the sound stage.
  • the diffuse sound may be spatially expanded to provide a more embracing sound stage without introducing spatial distortions or errors to non-diffuse/ direct sound.
  • the approach may be able to deliver both clearly localizable sounds as well as a very enveloping ambient sound. This may typically be achieved without the need for any user interaction.
  • first and second channels may specifically be a left front and right front channel of a stereo or a surround sound setup. In many embodiments the first and second channels may specifically be a left surround and right surround channel of a surround sound setup.
  • the upmixing which is applied to the first channel signal may also be applied to the second channel signal.
  • the directional radiation pattern from the two speaker arrangements may be substantially the same or may be different.
  • the beamwidth of the main lobe may in some embodiments be relatively narrow (say ⁇ 20°) or may e.g. in other embodiments be relatively broad (say ⁇ 120°).
  • the first speaker arrangement may have a directional radiation pattern which has two (or more) substantially equal lopes in which case either of these main lobes may comprise the direction to the listening position within their 3dB beamwidth.
  • the second speaker arrangement may have a directional radiation pattern which has two (or more) substantially equal lopes in which case neither of these main lobes comprises the direction to the listening position within their 3dB beamwidth. For example, for a second speaker arrangement being implemented by a bipolar speaker both the lobes will not include the direction to the listening position within their 3 dB beamwidth.
  • the first speaker arrangement may in use render audio to the listening position predominantly along non-reflected acoustic paths.
  • the first speaker may specifically be arranged such that more than half of the audio energy reaching the listening position from the first speaker arrangement within the first 20 ms after the first wavefront does so via one or more direct paths. Some of the sound within the 20 ms may possibly reach the listening position through reflected acoustic paths but more than half of the audio energy reaching the listening position from the first speaker arrangement in this time interval will in many embodiments and scenarios not be reflected. Sound outside the 20 ms time interval will typically be reverberant sound with few and weak spatial cues. The reverberant sound tends to be dependent only on room acoustics and not on the speaker setup and arrangements.
  • the second speaker arrangement may in use render audio to the listening position predominantly along reflected acoustic paths.
  • the second speaker may specifically be arranged such that more than half of the audio energy reaching the listening position from the second speaker arrangement within the first 20 ms after the first wavefront does not do so via one or more direct paths. Some of the sound within the 20 ms may possibly reach the listening position through direct non-reflected acoustic paths but more than half of the audio energy reaching the listening position from the second speaker arrangement in this time interval will in many embodiments and scenarios be reflected at least once. Typical reflections may be off the walls, ceiling or floor of the room in which the rendering system is located.
  • the second audio signal can correspond to a more diffuse sound than the first audio signal in that the second audio signal has a higher proportion of signal components for which the correlation measure indicates a lower correlation between the first channel signal and the second channel signal than for the first audio signal.
  • the second audio signal can correspond to a more uncorrected sound (between the first and second channel) than the first audio signal.
  • this reference may be considered with reference to the audio scene represented by the input multi-channel signal.
  • This audio scene may represent an audio environment with a number of spatially well defined (point like) sources as well as more diffuse sound components that are not spatially well-defined.
  • the second audio signal can correspond to more diffuse sound than the first audio signal in that it contains a higher proportion of the audio energy of the diffuse sound of the input multi- channel/ captured audio scene than does the first audio signal.
  • the first audio signal can correspond to less diffuse sound than the second audio signal by it containing a higher proportion of the audio energy of the spatially well-defined audio sources of the input multichannel/ captured audio scene than does the second audio signal.
  • diffuseness/ non-diffuseness when referring to the signals generally correspond to terms such as directional/non-directional, localizable/non-localizable and/or foreground/background.
  • the first audio signal may predominantly include sound components of the first channel signal corresponding to spatially specific audio sources (such as point-like sources) whereas the second audio signal may predominantly include sound components of the first channel signal corresponding to spatially non-specific ambient sound. Specifically, the second audio signal may predominantly reflect background sounds whereas the first audio signal may predominantly reflect specific foreground sound sources.
  • the audio renderer is arranged to divide the first channel signal into a plurality of time-frequency intervals; and the correlation estimator is arranged to generate a correlation value for each time- frequency interval; and the upmixer is arranged to generate the second audio signal by for each time frequency interval weighting a signal value of the first channel signal for the time frequency interval by a first weight being a monotonically decreasing function of a correlation value for the time- frequency interval.
  • This may provide a particularly advantageous approach.
  • it may provide an efficient separation of sound components which are highly correlated between channels and sound components that are not highly correlated.
  • the approach may allow an effective generation of a second audio signal which corresponds to diffuse sound components of the first audio channel.
  • the upmixer is further arranged to generate the first audio signal by for each time frequency interval weighting the signal value of the first channel signal for the time frequency interval by a second weight being a monotonically increasing function of the correlation value for the time- frequency interval.
  • This may provide a particularly advantageous approach.
  • it may provide an efficient separation of sound components which are highly correlated between channels and sound components that are not highly correlated.
  • the approach may allow an effective generation of a first audio signal which corresponds to non-diffuse sound components of the first audio channel, and a second audio signal which corresponds to diffuse sound components of the first audio channel.
  • the upmixer is further arranged to determine the weight in response to an energy difference estimate for the first channel signal and the second channel signal.
  • the approach may e.g. allow an improved separation into diffuse and non- diffuse sound. Specifically, it may provide improved consideration for spatially well defined (e.g. point like) sources that are planned to one of the first and second channels, i.e. for which the energy is predominantly located in one of the channels.
  • spatially well defined e.g. point like
  • the energy difference may be evaluated in individual time frequency intervals, over a group of time frequency intervals or over all frequencies.
  • the gain may be determined as a function of the energy difference and may specifically be a monotonically decreasing function of the energy difference.
  • the correlation estimator is arranged to determine the correlation value for the frequency interval in response to a frequency averaging of correlation values of a plurality of time frequency intervals.
  • This may provide improved performance and may in particular in many embodiments and for many signals may reduce distortion caused by the upmixing of the first channel signal.
  • the upmixer is further arranged to determine the weight in response to an audio content characteristic for the multichannel signal.
  • This may provide an improved user experience in many embodiments. For example, it may provide an improved adaptation of the rendering of diffuse and non-diffuse sound of the specific audio signal. For example, a sound stage more appropriate for the audio content may be generated.
  • the audio renderer may be arranged to modify a rendering property of the first audio signal independently of the second audio signal.
  • This may provide an improved user experience in many embodiments. For example, it may provide an improved adaptation of the rendering of diffuse and non-diffuse sound of the specific audio signal. For example, a sound stage more appropriate for the audio content may be generated.
  • the rendering property is an audio level for the first audio signal.
  • the audio renderer may be arranged to modify an audio level of the second audio signal independently of the first audio signal.
  • the rendering property is a spatial audio radiation pattern property.
  • the audio radiation pattern property may be a property of a beam pattern/shape, e.g. of a speaker array used with a dynamically variable beamformer.
  • the audio renderer may be arranged to modify a spatial audio radiation pattern property for the second audio signal independently of the first audio signal.
  • directional radiation pattern for the second speaker arrangement has a notch in the direction of the listening position.
  • the second speaker arrangement may specifically be an audio array controlled by a beamformer comprised in the second driver.
  • the adaptive beamformer may be arranged to (possibly dynamically) steer a null in the direction of the listening position.
  • the second speaker arrangement comprises a bipolar speaker arrangement. This may allow advantageous performance while maintaining a low complexity implementation.
  • the first speaker arrangement and the second speaker arrangement are comprised in one speaker enclosure.
  • the two speaker arrangements may be implemented by separate sets of one or more drive units angled in different directions.
  • the first and second speaker arrangements may be implemented by a single audio array driven by a different beamformer for each of the first and second audio signals with the beamformers generating beams in different directions.
  • the multi-channel audio signal is a spatial multi-channel signal having each channel associated with a nominal position of a spatial speaker configuration, and wherein the second speaker arrangement is located at a different position than the nominal position.
  • This may provide an improved user experience and a more encapsulating sound rendering in many embodiments. In particular, it may provide the perception of a larger sound stage while still maintaining the positions of point like audio sources.
  • the second driver is associated with an elevated speaker position.
  • This may provide an improved user experience and a more encapsulating sound rendering in many embodiments. In particular, it may provide the perception of a larger sound stage while still maintaining the positions of point like audio sources.
  • Fig. 1 illustrates an example of elements of an audio rendering system in accordance with some embodiments of the invention
  • Fig. 2 illustrates an example of elements of an audio renderer in accordance with some embodiments of the invention
  • FIG. 3 illustrates an example of a correlation measure between two channels of a multi-channel audio signal.
  • Fig. 1 illustrates an example of an audio rendering system in accordance with some embodiments of the invention.
  • an audio renderer 101 receives a multi- channel signal which in the specific example is a five channel spatial multi-channel signal.
  • the multi-channel signal may be a conventional five channel signal with spatial channels associated with loudspeakers positioned at specific spatial positions relative to a listening position 103.
  • a 5.1 home cinema system provides spatial sound via five loudspeakers being positioned with one speaker 105 directly in front of the listening position (the center channel), one speaker 107 to the front left of the listening position, one speaker 109 to the front right of the listening position, one speaker 111 to the rear left of the listening position, and one speaker 113 to the rear right of the listening position.
  • a non-spatial low frequency effects channel may be provided and rendered via a low frequency speaker (not shown).
  • the system of Fig. 1 may thus provide a spatial sound experience to a listener at the listening position 103.
  • the system of Fig. 1 is further arranged to synthesize additional channels from the received signals.
  • the system of Fig. 1 may decompose one channel into two channels and render the two channels from two different speaker arrangements.
  • the left front channel is decomposed into a first signal and a second signal, where the first signal drives a first speaker 109 which specifically may be positioned at the nominal position for the left front channel and the second signal drives a second speaker which may be collocated with the first speaker 109 or may be positioned elsewhere.
  • the right front channel is decomposed in a similar way and thus an additional speaker arrangement 117 is used to render the additional signal.
  • the signal in each of the two front side channels is thus divided into two different signals.
  • one of the generated signals predominantly corresponds to non-diffuse sound, such as sound from specific (point like) sound sources, whereas the other signal predominantly corresponds to more diffuse sound.
  • differentiation and the decomposition is based on an evaluation of the correlation between different channels of the multi-channel audio signal.
  • point-like sources tend to exhibit a high degree of correlation between channels whereas diffuse sounds such as those originating from e.g. reverberation effects, non-directional noise etc tend not to exhibit a high degree of correlation.
  • the individual characteristics of the two channels are further emphasized by the rendering of the different signals being different.
  • the non-diffuse signal is to a large extent rendered via direct acoustic paths whereas the diffuse signal is to a large extent rendered via indirect paths.
  • the system may specifically apply a blind decomposition algorithm which seeks to differentiate between ambient and more diffuse background sound and specific spatially well-defined foreground sound sources.
  • an audio scene in a movie may often consist of sound sources that are at the foreground to the listener (like dialogue and some effects) and sound elements that are at a larger distance or at the background (environmental sounds and sometimes background music).
  • the latter type of sound will typically be diffuse sound with few specific spatial cues.
  • the blind decomposition of the content in this way would be extremely difficult without additional cues.
  • the original mixture has already been created in such a way that the foreground and background sound elements are mixed in different ways.
  • foreground sound elements appear typically in only one or two loudspeakers in which case they have a large signal-level cross-correlation at zero time-lag.
  • background sound elements are typically placed in two or more loudspeakers and they are typically weakly zero-lag cross-correlated between pairs of channels.
  • Some foreground sounds which are panned predominantly to one channel may also exhibit a low cross-correlation but as will be described later, such scenarios may explicitly be detected and compensated for.
  • the two signals can be rendered by speaker arrangements having different directional radiation patterns.
  • the speaker arrangement rendering the foreground signals may be aimed directly at the listening position 103, i.e. the listening position may fall within the (3dB) beamwidth of the mainlobe of the speaker arrangement.
  • the speaker arrangement rendering the background signals may be aimed away from the listening position 103.
  • the listening position may thus be outside the (3dB) beamwidth of the mainlobe.
  • Such an arrangement may thus ensure that the proportion of sound rendered via direct acoustic paths relative to the proportion of sound rendered via reflected acoustic paths is much higher for the rendering of foreground objects than for the rendering of background objects. Thus, the relative diffuseness of the rendered sound is increased for background objects relative to the foreground objects.
  • the foreground signals can be reproduced by a speaker arrangement which for a non-reverberant time interval (of 20 ms) predominantly renders the signal to the listening position 103 via direct acoustic paths thereby providing strong spatial cues resulting in clearly localizable sound images.
  • the background signals may be rendered by a speaker
  • Fig. 2 illustrates an example of some elements of the system of Fig. 1 which are related to the generation of two output channels from one input channel. Specifically, the figure may be considered to illustrate elements for the left front channel of Fig. 1 but it will be appreciated that the approach is equally applicable to the right front channel and indeed to any audio signal which is upmixed to two output channels that are then differently rendered.
  • Fig. 2 illustrates the audio renderer 101 comprising a receiver 201 that is arranged to receive the multi-channel signal.
  • the receiver 201 is coupled to an upmixer 203 which is fed the signal of one of the signals of the multi-channel signal, and in the specific example it is fed the left front channel.
  • the upmixer 203 is arranged to upmix the received signal to generate two output signals.
  • the second signal comprises a higher proportion of diffuse sound than the second signal.
  • the upmixer 203 may divide the input signal into sound components that predominantly correspond to diffuse or non-spatially well-defined sound sources, and sound components that are not diffuse but typically are spatially relatively well defined.
  • the first signal may typically predominantly correspond to specific foreground elements whereas the second signal may typically correspond to background sound.
  • the two signals will henceforth be referred to as a foreground signal and a background signal.
  • the decomposition into the foreground and background signals are performed by considering the correlation between two channels of the multi-channel signal.
  • the approach may specifically exploit that diffuse/ background signals tend to be generated to have low correlation between different channels of the multi-channel signal whereas point-like/ specific foreground objects tend to have a high correlation.
  • the upmixer 203 may thus decompose the signal by seeking to direct sound components with high correlation to the foreground signal and sound components with low correlation to the background signal.
  • the foreground signal may comprise a higher concentration of correlated sound components than the background signal.
  • the foreground signal is fed to a first driver 205 which is coupled to the upmixer 203 and an external speaker arrangement 107 (henceforth referred to as the foreground speaker 107) which may comprise one or more speaker drivers/ audio transducers.
  • the background signal is fed to a second driver 207 which is coupled to the upmixer 203 and an external speaker arrangement 115 (henceforth referred to as the background speaker 115) which may comprise one or more speaker drivers/ audio transducers.
  • the two generated signals are rendered independently using different speaker arrangements (also for brevity referred to as speakers although it will be appreciated that these may comprise a plurality of speaker drivers, and may indeed share some speaker drivers e.g. using an audio array and beamforming to render the channels).
  • speaker arrangements also for brevity referred to as speakers although it will be appreciated that these may comprise a plurality of speaker drivers, and may indeed share some speaker drivers e.g. using an audio array and beamforming to render the channels).
  • the individual speakers are arranged to provide a rendering which is particularly suitable for the specific type of audio signal rendered.
  • the characteristics of the speakers are such that it provides a particularly advantageous rendering for the individual characteristics of the two generated signals.
  • both the foreground speaker arrangement 107 and the background speaker arrangement 115 are directional speakers and thus have a directional radiation pattern (e.g. given as a relative gain as a function of angle of radiation).
  • the directional radiation pattern has a mainlobe for which the maximum radiation level (the maximum gain) is achieved.
  • the beamwidth of such a mainlobe may be determined as the 3dB beamwidth given as the width of the beam between the two points at which the radiation (power) level (the gain) has dropped to 3dB lower than the maximum radiation level (gain).
  • the radiation pattern may exhibit a plurality of identical lobes (i.e. there may be more than one mainlobe).
  • the foreground speaker arrangement 107 is arranged such that the listening position 103 falls within the 3dB beamwidth of the mainlobe (or of any of the mainlobes in case there are more than one).
  • the background speaker arrangement 115 is arranged such that the listening position 103 does not fall within the 3dB beamwidth of the mainlobe (or of all the mainlobes in case there are more than one). This arrangement may specifically allow the rendering of the foreground signal being
  • the foreground speaker 107 is arranged to render audio to a listening position which within a non-reverberant time interval of 20 ms is predominantly rendered along non-reflected acoustic paths from the foreground speaker 107 to the listening position 107.
  • a direct rendering provides strong spatial cues and provides a listener with spatial cues that allows for sound components rendered from the foreground speaker 107 to be perceived to originate from the position of the foreground speaker 107.
  • This, together with corresponding sound components from the other spatial channels (and especially from the front right and the center channels), provide a panning effect that allows specific spatially well-defined audio elements to be positioned in the sound scene and to be perceived as sound sources with specific well defined positions.
  • the background speaker 115 is in contrast arranged to render audio to a listening position 103 which within a non-reverberant time interval of 20 ms is
  • the background speaker 115 predominantly rendered along reflected acoustic paths from the background speaker 115 to the listening position 103.
  • at least half of the early audio energy (within the 20 ms of the first wavefront) from the background speaker 115 reaches the listening position 103 via non-direct, reflected rendering.
  • at least 75% or even 90% of the sound energy may be via reflected paths.
  • the reflections may occur of walls, floor, ceiling, obstacles etc in the room in which the system is located.
  • Such an indirect rendering results in the rendered audio being spread in both time and space, and it will reduce the amount of spatial cues relating to the speaker position which is provided to a listener.
  • the listener may instead perceive sound that is spread and with a more pronounced diffuse characteristic.
  • the use of reflected sound enhances the diffuse nature of the background signal which corresponds to the more diffuse background or ambient sounds.
  • Such diffuse sound is particularly suitable to provide the listener with a perception of a larger and more encapsulating sound scene without introducing e.g. the perception of phantom or moved audio sources.
  • a significant part of the rendered energy reaches the listening position as reverberant signal components.
  • Such a reverberation tail of the acoustic transfer function from a speaker to a listening position may be relatively long and difficult to estimate.
  • the reverberant propagation tends to be independent of the specific speaker setup and is generally predominantly dependent on room characteristics.
  • the reverberant tail provides very limited spatial cues to the listener.
  • the differentiation between the renderings of the two speaker arrangements is used to provide a differentiation in the spatial perception. Accordingly, they are arranged to provide very different rendering for the initial non-reverberant time interval and the characteristics for the reverberant tail are less significant. Therefore, the speaker
  • the two speaker arrangements are arranged such that the audio reaching the listening position 103 within 20 ms of the first wave front does so via direct paths for the first speaker arrangement 101.
  • the first 20 ms from the earliest non-zero value of the acoustic transfer function (i.e. from the first wave front reaches the listening position) from the speaker arrangements to the listening position 103 is for the foreground speaker 107 predominantly a result of direct acoustic paths and for the background speaker 115 predominantly a result of reflected paths.
  • references to the differentiations in rendering and the difference between the rendering of the foreground speaker 107 and the background speaker 115 may for brevity not explicitly refer to the characteristics being for this 20 ms time interval, but it will be appreciated that the references to e.g. rendering being predominantly via direct or indirect acoustic paths are to be considered within this time interval.
  • the upmixer 203 is arranged to generate the foreground and background signals based on an evaluation of the correlation of the channel being upmixed (in the specific example the left front channel) with another channel. Specifically, a correlation measure which is indicative of the correlation between the channel being upmixed and another channel is used by the upmixer to synthesize the new signals.
  • the audio renderer 101 comprises a correlation estimator 213 which is arranged to generate a correlation measure for the signal of the channel being upmixed and the signal of another channel.
  • the correlation measure may typically and in many scenarios advantageously be indicative of the correlation of the left front channel to the right front channel.
  • the correlation measure may typically and in many scenarios advantageously be indicative of the correlation of the left surround channel to the right surround channel.
  • the correlations are of course equally appropriate for the right front and right surround channels respectively.
  • the correlation estimator 213 is arranged to generate the correlation measure by performing a direct correlation.
  • the correlation measure may comprise a specific correlation value for each of a plurality of time frequency intervals, also referred to as time- frequency tiles. Indeed, the upmixing of the signal may be performed in time- frequency tiles and the correlation measure may provide a correlation value for each time- frequency tile.
  • the resolution of the correlation measure may be lower than that of the time- frequency tiles of the upmixing.
  • a correlation value may be provided for each of a number of perceptual significance bands, such as for each of a number of ERB bands.
  • Each perceptual significance band may cover a plurality of time- frequency tiles.
  • the correlation measure may be fed to the upmixer 203 which can proceed to determine gains for respectively the foreground and the background signal.
  • the input signal may be segmented and converted to the frequency domain.
  • the upmixer 203 may generate a foreground signal value by multiplying it by a foreground gain derived from the correlation value for the corresponding time- frequency tile.
  • the foreground gain may increase for increasing correlation.
  • a frequency domain signal is generated that comprises a high weighting of the correlated components of the input signal.
  • the upmixer 203 may generate a background signal value by multiplying it by a background gain derived from the correlation value for the corresponding time- frequency tile.
  • the background gain may be decrease for increasing correlation.
  • a frequency domain signal is generated that comprises a low weighting of the correlated components of the input signal.
  • the two generated frequency signals may then be converted back to the time domain to provide the background and foreground signals.
  • the upmixer 203 may specifically determine the foreground gain and the background gain to exactly or approximately maintain the overall energy level of the signals (specifically the sum, or the sum of the square, of the gains may be set to one).
  • the upmixer 203 may furthermore be arranged to provide a frequency domain smoothing of the gains which may improve the perceived sound quality.
  • both the left front and the right front channels are upmixed and thus the upmixing is applied to a stereo signal
  • the upmixed vector signal is given by
  • ⁇ ( ⁇ , ⁇ ) G(n, co)X(n, co) where G( «, ⁇ ) is a matrix operation.
  • the filter matrix can in the specific example be written in the following form: g (n, (o) 0
  • This matrix does not mix left and right channels (zeroes in the matrix). This is a design choice, and it will be appreciated that it is also possible to design algorithms where the cross-channel terms are non-zero resulting in mixing between the two sides. This may typically be more interesting for the synthesis of the background channels than for the synthesis of the foreground channels.
  • the gains of the matrix are determined from the correlation measure.
  • the weights for the foreground signals are determined as monotonically increasing functions of the correlation measure (and specifically of the correlation value in that time frequency tile).
  • the allocation of the signal energy of a specific time frequency tile into the foreground signal increases the more the two spatial channels are correlated.
  • the gains may also depend on other parameters and considerations but that the relationship to the correlation value will be monotonically increasing.
  • the weights for the background signals are determined as monotonically decreasing functions of the correlation measure (and specifically of the correlation value in that time frequency tile).
  • the allocation of the signal energy of a specific time frequency tile into the background signal increases the less the two spatial channels are correlated, i.e. the more it corresponds to diffuse sound.
  • the gains may also depend on other parameters and considerations but that the relationship to the correlation value will be monotonically decreasing.
  • the upmixer 203 decomposes the side front signals into signal components that are correlated and signal components that are not correlated, and thus typically into diffuse ambient sound and non-diffuse foreground sound.
  • the correlation estimator 213 determines the correlation values which in the specific example is between the two front channels.
  • the correlation coefficient can be defined as:
  • ⁇ .. .> denotes the computation of an expected value of the inner product of the two data sets over the variable n .
  • the correlation value for each time- frequency tile may then be determined as: Often resulting from the (frequency) bin-by-bin operation, any highly variable function in the frequency domain can create a significant amount of audible artifacts when applied as a gain function for audio signal processing.
  • the black solid line in Fig. 3 shows an example of such weighting (gain) functions, which actually is the correlation values obtained according to the equations above. Although each value on this curve may represent the desired functionality of the weighting function, an additional averaging process in the frequency direction may improve audio quality substantially in many scenarios.
  • the correlation estimator 213 is therefore furthermore arranged to determine the correlation value for a given time frequency interval in response to a (weighted) frequency averaging of correlation values of a plurality of time frequency intervals.
  • a spectral smoothing can be performed.
  • S[-] indicates a suitable frequency smoothing function.
  • a triangular or square smoothing function may be applied.
  • the smoothing function S may simply determine the average of the unsmoothed correlation value for the current time frequency tile and the N surrounding (in the frequency domain) unsmoothed correlation values.
  • the correlation estimator 213 may optionally determine the gain in response to an energy difference estimate for the channels.
  • an important special case that may be considered is when a strong spatially well-defined sound source is concentrated in one speaker, e.g. when the sound source is hard-panned to the left or right.
  • the correlation coefficient will also become small, which would indicate to the system that the corresponding time- frequency region is likely to be ambient diffuse sounds. This is usually not desired as the extreme side panned content is typically intended to be on the extreme sides in the stereo image rather than rendered diffusely.
  • the system may specifically seek to address such an issue. This is in the example done by adapting the gains in response to an energy difference between the channels.
  • an additional weight function h(n, co) for the gains may be determined on the basis of estimates of signal energy differences between the two channels.
  • h(n, a>) is positive in spectral regions where channel 1 dominates and negative in areas where the other channel has more energy.
  • the system thus separates out components that are likely to correspond to diffuse ambient/background sounds and components that are likely to correspond to non- diffuse foreground sources thereby providing an upmixing to two distinct channels with characteristic properties.
  • the audio renderer 101 may be arranged to individually adapt properties of the rendering for the two channels.
  • the audio renderer 101 can change or set a rendering property for one of the signals independently of the setting for the other signal.
  • the rendering signal processing may specifically be adapted by means of a user control e.g. for controlling the applicable limits of the degree of diffuseness.
  • the audio renderer 101 can set the audio level for one of the signals independently of the other signal.
  • the volume for the background signal relative to the volume for the foreground signal may be modified and set to provide a desirable audio experience.
  • the volumes of the background and foreground may be set individually for the two front side signals. This may provide an improved user experience in many scenarios. For example, it may allow emphasis of the dialogue relative to the background sound thereby e.g. aiding users with hearing difficulties.
  • the system may change the spatial rendering characteristics individually for the two signals.
  • the system can render the individual types of sound differently, and especially can render the foreground sound such that it provides strong spatial cues relating to the position of the speaker whereas the background sound is rendered via reflected paths thereby not providing strong spatial cues about the position of the speaker rendering the sound.
  • the radiations pattern (e.g. the beam pattern) for one of the speakers 107, 115 may be dynamically adaptable.
  • one of the speakers 107, 115 may be implemented using a speaker array with a dynamic adaptable beamformer.
  • the same audio array may together with different beamformers render both the background signal and the foreground signal, i.e. both speaker arrangements 107, 115 may be implemented by the same audio array but using different beamform parameters to provide rendering in different directions.
  • the system may individually steer the audio rendering in different directions for the two signals. For example, the system may track the position of a listener e.g. using a video based head tracking system. The beamform parameters may then be individually adapted for the two signals based on the position of the user. E.g. for the foreground signal the beamform weights can be set to direct a maximum of the beam-shape in the direction of the listening position whereas for the background signal the beamform weights can be set to direct a null in the direction of the listening position.
  • the speaker arrangements may be implemented as one or two audio arrays driven by two different beamformers (or equivalently the same physical beamforming functionality using different beamform weights).
  • the beamform weights may in some embodiments be fixed thereby providing a fixed radiation pattern.
  • the audio array may in such cases be angled to provide strong direct paths to the listening position for the foreground signal but not for the background signal. Rather, the array may be angled to provide a notch (and typically a null) of the beam-pattern in the direction of the listening position.
  • the speaker used for the foreground signal may be a conventional speaker driver directed towards the listening position.
  • the background speaker arrangement may be a conventional speaker driver which is directed away from the listening position and typically towards a wall for providing suitable reflections.
  • the two speakers 115, 107 can be comprised in a single speaker enclosure with the arrangements being such that the radiation patterns are in different directions.
  • a foreground speaker may be positioned on a front-firing
  • the foreground speaker When the speaker enclosure is positioned in the nominal position and angled towards the listening position, the foreground speaker will predominantly render the audio along direct paths whereas the background speaker will typically render the audio via reflections of e.g. a wall to the side of the speaker.
  • the background speaker arrangement 115 may be implemented by a bipolar speaker arrangement.
  • two drivers may be fed the same drive signal but with a 180° phase difference and with the two drivers being directed in opposite directions.
  • This approach will generate a strong sound radiation in two opposite directions with a null in-between.
  • the null can be directed towards the listening position.
  • This arrangement provides low complexity, and thus low cost, implementation yet can provide a strong rendering of the background signal in several directions thereby providing many different reflections. Furthermore, the direct path audio rendering can be minimized.
  • a diffuse rendering of the background signal can be achieved via a low cost implementation.
  • the approach may be particularly suitable for implementations in a single enclosure with the two drivers of the bipolar arrangement being arranged in a side-firing configuration with a third driver used for rendering the foreground signal being arranged in a front-firing configuration.
  • both the background and the foreground signals may be rendered from the same position, and indeed from the nominal or reference position associated with the spatial audio channel from which they are generated. Such approaches may specifically use a single speaker enclosure comprising both speaker arrangements. However, in other embodiments, at least one of the generated signals may be rendered from a different position. Specifically, in many embodiments, the foreground signal may be rendered from the reference or nominal position of the channel that was upmixed. This ensures that the positions of the foreground objects in the audio stage are not modified. However, the background signal may be rendered from a different position than the foreground signal and specifically from another position than the nominal position of the channel being upmixed. This may provide an expanded sound stage and may in particular provide a perception of a substantially larger sound stage.
  • the background speakers may be rendered from elevated speakers thereby providing a sound stage which extends outside the horizontal plane normally associated with rendering configurations.
  • a similar effect may be achieved using (at least partially) upfiring speaker drivers for the background signal where the upfiring speakers are provided in the same enclosure as the speaker driver(s) for the foreground signal.
  • the approach can be adapted to generate appropriate signals for such elevated speakers.
  • the available media such as discs or broadcasts do not contain dedicated height signals.
  • the described upmixing algorithm may be used.
  • Existing solutions often generate signals that are not uncorrected from the other channels, thereby potentially elevating the complete sonic image including the principal sound sources. This is not favorable since the desired location of these sources is in most cases on the horizontal plane and the rendering from the elevated positions will result in a position offset from the horizontal plane being introduced.
  • Other solutions avoid this issue by generating height signals with a rather low audio level. In both cases, the possible advantages of elevated loudspeakers are not fully used.
  • the described approach can be used to extract audio signal components that predominantly correspond to more diffuse background sound.
  • the corresponding signal can then be reproduced through e.g. elevated loudspeakers, thereby increasing the sonic envelopment and sense of realism, while not introducing disturbing artifacts such as localization shifts.
  • the described approach may be applied to a plurality of the channels/ channel set.
  • the described approach for the front left and right channels may also be applied to the surround left and right channels.
  • the system may accept five input signals such as the spatial channels of a 5.1 surround sound audio, and may output nine loudspeaker signals, which are center, directional left/right/surround-left/surround-right, and diffuse left/right/surround-left/surround-right.
  • the decomposed signals may be recomposed for at least one of the signals.
  • the output signal for the speaker at the nominal position may be generated as a combination of the foreground signal and of the background signal. This recombination may allow the diffuse background sounds to be rendered not only from the second speakers (e.g. elevated speakers) but also from the original positions. However, typically the relative level of the background signal components will be reduced with respect to the original signal to compensate for the rendering being along direct paths and for the additional rendering of background sounds which is provided by the additional speakers.
  • the upmixer 203 is further arranged to determine the gains used to decompose the input signal into the background signal and foreground signal in response to an audio content characteristic for the received multi-channel signal.
  • the balance between the direct and ambient channels can be adjusted and this may specifically be used to automatically adapt the processing depending on the audio content.
  • the audio content may for example be characterized by metadata describing the content.
  • metadata may be provided to describe whether the audio is sound from e.g. a football game (having few foreground sources with significant diffuse background sound (ambient sound of the crowd)), from a discussion program (only few foreground sound sources with typically very little background sound) etc.
  • the gains may be adjusted depending on such values. For example, for each content category a scale factor may be stored which scales the gains for the background and foreground decomposition (in opposite directions).
  • the adaptation may be in response to a characteristic of the audio signal, such as an averaged frequency response, relative signal energies of all multichannels etc.
  • the invention can be implemented in any suitable form including hardware, software, firmware or any combination of these.
  • the invention may optionally be
  • an embodiment of the invention may be physically, functionally and logically implemented in any suitable way. Indeed the functionality may be implemented in a single unit, in a plurality of units or as part of other functional units. As such, the invention may be implemented in a single unit or may be physically and functionally distributed between different units, circuits and processors.

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Multimedia (AREA)
  • Stereophonic System (AREA)
  • Circuit For Audible Band Transducer (AREA)

Abstract

L'invention concerne un système de rendu audio qui comprend un premier agencement de haut-parleurs (107) de rendu audio vers une position d'écoute (103) principalement le long de trajets acoustiques non réfléchis et un deuxième agencement de haut-parleurs (115) de rendu audio principalement le long des trajets réfléchis. Un dispositif de rendu audio (101) comprend un mélangeur élévateur (203) destiné à mélanger et élever un signal de canal d'un signal audio multiplex vers un premier signal audio et un deuxième signal audio correspondant à un son plus diffus que celui du premier signal audio. Le mélange et l'élévation s'effectuent en réponse à la mesure de corrélation pour deux canaux du signal multiplex. La mesure de corrélation est produite par un estimateur de corrélation (213). Les moteurs de haut-parleur (205, 207) entraînent les premier et deuxième agencements de haut-parleurs (107, 115) des premier et deuxième signaux audio respectivement. Cette approche peut offrir une expérience audio plus enveloppante sans introduire d'erreur de position pour des sources sonores spécifiques.
EP13710012.9A 2012-01-23 2013-01-14 Système de rendu audio et procédé correspondant Withdrawn EP2807833A2 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US201261589438P 2012-01-23 2012-01-23
PCT/IB2013/050331 WO2013111034A2 (fr) 2012-01-23 2013-01-14 Système de rendu audio et procédé correspondant

Publications (1)

Publication Number Publication Date
EP2807833A2 true EP2807833A2 (fr) 2014-12-03

Family

ID=47891792

Family Applications (1)

Application Number Title Priority Date Filing Date
EP13710012.9A Withdrawn EP2807833A2 (fr) 2012-01-23 2013-01-14 Système de rendu audio et procédé correspondant

Country Status (4)

Country Link
EP (1) EP2807833A2 (fr)
JP (1) JP6078556B2 (fr)
CN (1) CN104041079A (fr)
WO (1) WO2013111034A2 (fr)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9743211B2 (en) 2013-03-19 2017-08-22 Koninklijke Philips N.V. Method and apparatus for determining a position of a microphone
CH709272A2 (de) * 2014-02-28 2015-08-28 Stormingswiss S Rl C O Fidacor S Rl Autonome Residualbestimmung und Gewinnung von residualarmen Zusatzsignalen.
WO2015173422A1 (fr) * 2014-05-15 2015-11-19 Stormingswiss Sàrl Procédé et dispositif pour la réalisation sans résiduelle d'un mixage élévateur à partir d'un mixage réducteur
US10225676B2 (en) * 2015-02-06 2019-03-05 Dolby Laboratories Licensing Corporation Hybrid, priority-based rendering system and method for adaptive audio
DE102015008000A1 (de) * 2015-06-24 2016-12-29 Saalakustik.De Gmbh Verfahren zur Schallwiedergabe in Reflexionsumgebungen, insbesondere in Hörräumen
KR102319880B1 (ko) * 2016-04-12 2021-11-02 코닌클리케 필립스 엔.브이. 포커스 거리에 가까운 사운드 소스들을 강조하는 공간 오디오 처리
US10869151B2 (en) 2016-05-31 2020-12-15 Sharp Kabushiki Kaisha Speaker system, audio signal rendering apparatus, and program
GB2574239A (en) * 2018-05-31 2019-12-04 Nokia Technologies Oy Signalling of spatial audio parameters
CN112005210A (zh) * 2018-08-30 2020-11-27 惠普发展公司,有限责任合伙企业 多通道源音频的空间特性
GB201818959D0 (en) * 2018-11-21 2019-01-09 Nokia Technologies Oy Ambience audio representation and associated rendering
WO2020178256A1 (fr) * 2019-03-04 2020-09-10 A-Volute Appareil et procédé d'analyse audio
US11943600B2 (en) * 2019-05-03 2024-03-26 Dolby Laboratories Licensing Corporation Rendering audio objects with multiple types of renderers
EP3761672B1 (fr) 2019-07-02 2023-04-05 Dolby International AB Utilisation de métadonnées pour agréger des opérations de traitement de signal
GB2587371A (en) * 2019-09-25 2021-03-31 Nokia Technologies Oy Presentation of premixed content in 6 degree of freedom scenes
US11373662B2 (en) * 2020-11-03 2022-06-28 Bose Corporation Audio system height channel up-mixing
CN114040319B (zh) * 2021-11-17 2023-11-14 青岛海信移动通信技术有限公司 一种终端设备外放音质优化方法、装置、设备和介质

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4837825A (en) * 1987-02-28 1989-06-06 Shivers Clarence L Passive ambience recovery system for the reproduction of sound
US7231060B2 (en) * 1997-08-26 2007-06-12 Color Kinetics Incorporated Systems and methods of generating control signals
US7254239B2 (en) * 2001-02-09 2007-08-07 Thx Ltd. Sound system and method of sound reproduction
GB0219245D0 (en) * 2002-08-17 2002-09-25 Kh Technology Corp Combination loudspeaker unit
JP4418774B2 (ja) * 2005-05-13 2010-02-24 アルパイン株式会社 オーディオ装置およびサラウンド音生成方法
EP1989920B1 (fr) * 2006-02-21 2010-01-20 Koninklijke Philips Electronics N.V. Codage et décodage audio
CN201328191Y (zh) * 2008-12-18 2009-10-14 固昌通讯股份有限公司 多声道音源转接盒与多声道音响***
KR101485462B1 (ko) * 2009-01-16 2015-01-22 삼성전자주식회사 후방향 오디오 채널의 적응적 리마스터링 장치 및 방법
US20120045065A1 (en) * 2009-04-17 2012-02-23 Pioneer Corporation Surround signal generating device, surround signal generating method and surround signal generating program
RU2551792C2 (ru) * 2010-06-02 2015-05-27 Конинклейке Филипс Электроникс Н.В. Система и способ для обработки звука

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
None *
See also references of WO2013111034A2 *

Also Published As

Publication number Publication date
JP2015508245A (ja) 2015-03-16
WO2013111034A2 (fr) 2013-08-01
JP6078556B2 (ja) 2017-02-08
WO2013111034A3 (fr) 2014-01-23
CN104041079A (zh) 2014-09-10

Similar Documents

Publication Publication Date Title
WO2013111034A2 (fr) Système de rendu audio et procédé correspondant
JP6950014B2 (ja) 2dセットアップを使用したオーディオ再生のためのアンビソニックス・オーディオ音場表現を復号する方法および装置
JP6186436B2 (ja) 個々に指定可能なドライバへの上方混合されたコンテンツの反射されたおよび直接的なレンダリング
US9955262B2 (en) Device and method for driving a sound system and sound system
US8699731B2 (en) Apparatus and method for generating a low-frequency channel
US8559661B2 (en) Sound system and method of operation therefor
JP5992409B2 (ja) 音再生のためのシステム及び方法
KR101546514B1 (ko) 오디오 시스템 및 그의 동작 방법
US20080298597A1 (en) Spatial Sound Zooming
CN108781341B (zh) 音响处理方法及音响处理装置
EP1761110A1 (fr) Méthode pour générer de l'audio multi-canaux à partir de signaux stéréo
JP4620468B2 (ja) オーディオ信号を再生するためのオーディオ再生システムおよび方法
JP2016509429A (ja) オーディオ装置及びそのための方法
JP2003510924A (ja) 音響指向方法および装置
JP4977720B2 (ja) Wfsシステムのシミュレーション及び音響特性の補償のための装置及び方法
JPH11504176A (ja) 3次元サウンドイメージを生成するための音響オーディオシステム
US20110069850A1 (en) Audio reproduction system comprising narrow and wide directivity loudspeakers
CN107873135B (zh) 声音***
US11924623B2 (en) Object-based audio spatializer
US11665498B2 (en) Object-based audio spatializer
CN111971978B (zh) 用于在多通道音频再现***中应用基于时间的效果的方法和***
JP2023548570A (ja) オーディオシステムの高さチャネルアップミキシング
KR20180024612A (ko) 오디오 신호 처리 방법 및 장치

Legal Events

Date Code Title Description
PUAI Public reference made under article 153(3) epc to a published international application that has entered the european phase

Free format text: ORIGINAL CODE: 0009012

17P Request for examination filed

Effective date: 20140624

AK Designated contracting states

Kind code of ref document: A2

Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR

RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: WOOX INNOVATIONS BELGIUM NV

DAX Request for extension of the european patent (deleted)
RAP1 Party data changed (applicant data changed or rights of an application transferred)

Owner name: GIBSON INNOVATIONS BELGIUM NV

17Q First examination report despatched

Effective date: 20180213

STAA Information on the status of an ep patent application or granted ep patent

Free format text: STATUS: THE APPLICATION IS DEEMED TO BE WITHDRAWN

18D Application deemed to be withdrawn

Effective date: 20180626