EP2255359A1 - Vorrichtung und verfahren zur akustischen anzeige - Google Patents
Vorrichtung und verfahren zur akustischen anzeigeInfo
- Publication number
- EP2255359A1 EP2255359A1 EP09721864A EP09721864A EP2255359A1 EP 2255359 A1 EP2255359 A1 EP 2255359A1 EP 09721864 A EP09721864 A EP 09721864A EP 09721864 A EP09721864 A EP 09721864A EP 2255359 A1 EP2255359 A1 EP 2255359A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- signal
- loudspeaker
- signals
- acoustic
- objects
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims description 17
- 230000015572 biosynthetic process Effects 0.000 claims description 17
- 238000003786 synthesis reaction Methods 0.000 claims description 17
- 230000005236 sound signal Effects 0.000 claims description 13
- 238000004590 computer program Methods 0.000 claims description 6
- 230000003068 static effect Effects 0.000 claims description 3
- 230000008901 benefit Effects 0.000 description 5
- 230000000007 visual effect Effects 0.000 description 5
- 230000003247 decreasing effect Effects 0.000 description 4
- 238000003491 array Methods 0.000 description 3
- 230000006870 function Effects 0.000 description 3
- 238000013507 mapping Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- XLYOFNOQVPJJNP-UHFFFAOYSA-N water Substances O XLYOFNOQVPJJNP-UHFFFAOYSA-N 0.000 description 3
- 230000008859 change Effects 0.000 description 2
- 238000001514 detection method Methods 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 230000004913 activation Effects 0.000 description 1
- 238000013459 approach Methods 0.000 description 1
- 230000003190 augmentative effect Effects 0.000 description 1
- 238000010009 beating Methods 0.000 description 1
- 238000012512 characterization method Methods 0.000 description 1
- 238000004040 coloring Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 230000003111 delayed effect Effects 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 231100001261 hazardous Toxicity 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000004091 panning Methods 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 238000012913 prioritisation Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000009877 rendering Methods 0.000 description 1
- 230000001020 rhythmical effect Effects 0.000 description 1
- 230000000630 rising effect Effects 0.000 description 1
- 230000003595 spectral effect Effects 0.000 description 1
- 238000012876 topography Methods 0.000 description 1
- 230000007704 transition Effects 0.000 description 1
- 238000010626 work up procedure Methods 0.000 description 1
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S7/00—Indicating arrangements; Control arrangements, e.g. balance control
- H04S7/30—Control circuits for electronic adaptation of the sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/11—Positioning of individual sound objects, e.g. moving airplane, within a sound field
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2400/00—Details of stereophonic systems covered by H04S but not provided for in its groups
- H04S2400/13—Aspects of volume control, not necessarily automatic, in stereophonic sound systems
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04S—STEREOPHONIC SYSTEMS
- H04S2420/00—Techniques used stereophonic systems covered by H04S but not provided for in its groups
- H04S2420/13—Application of wave-field synthesis in stereophonic audio systems
Definitions
- the present invention relates to an apparatus and method for acoustically displaying a position of an object in a playback room.
- Exemplary embodiments include in particular acoustic displays for use on ships.
- the present invention has the object to provide an apparatus and a method which displays a position of an object acoustically. This object is achieved by a device according to claim 1 or claim 18 and a method according to claim 20.
- the core idea of the present invention is that a plurality of loudspeakers is spatially arranged so differently in a reproduction room that different positions can be acoustically represented by different activation of the loudspeakers.
- a signal allocation device is designed to allocate an acoustic signal to the object
- a loudspeaker drive device is designed to determine one or more loudspeaker signals for the multiplicity of loudspeakers.
- the one or more loudspeaker signals are arranged to indicate the position of the object, wherein the one or more loudspeaker signals are based on the acoustic signal associated with the object by the signal assigning means.
- the one or more loudspeaker signals are determined so that when the one or more loudspeaker signals are reproduced, the position of the object in the playback room is displayed acoustically.
- Embodiments of the present invention also relate to how sensor signals can be displayed more easily by means of intelligent acoustic displays and thus both the security can be improved and the running costs can be reduced.
- Another idea of the present invention is based on the fact that an essential part of the information in many detectors is a location.
- a detector for example, a radar, a depth sounder, nautical charts or weather maps come into consideration and the location refers to, for example, a direction as well as a distance to the object.
- a sound field is generated, for example by means of several speakers, which encodes this information as precisely as possible in a natural way.
- Wavefield Synthesis in this system, for example, the loudspeakers are at a constant distance and the individual signals for the loudspeakers are calculated according to the well-known WFS algorithms. Objects from a radar signal are reproduced as acoustic objects in the corresponding direction and distance. The objects thus appear as virtual sound sources and can be localized by a listener. For example, all persons on the bridge can perceive the objects in the same place. It is also possible that not only a single object but also several objects are displayed acoustically at the same time, wherein each object, for example, another or optionally also a same acoustic signal can be assigned.
- WFS Wavefield Synthesis
- the workup includes on the one hand the detection of moving objects, such as ships and aircraft, and also the detection of static objects, such as the coastline, buoys or islands.
- moving objects such as ships and aircraft
- static objects such as the coastline, buoys or islands.
- the audio signal can optionally be converted into an audio signal by means of a text-to-speech identification, so that the text signal of the transponder becomes audible.
- Such objects are z. B. determines buoys or beacons, whose identifying information appear, for example, on the radar as text.
- Objects can still be classified according to their hazard potential. For example, objects that come closer (from the front or faster from the back) or cross the ship's path of movement may be classified as more dangerous than objects that run parallel to the ship or are moving away from the ship. Objects that are farther away are generally considered less dangerous than those that are near or approaching at a high relative speed.
- a different identifier tone with the identification tone for example in pitch or in the pulse repetition frequency and increase as the danger increases.
- a higher tone may mean greater danger or increasing volume may imply an increasing danger.
- a faster beating clock pulse may mean a rising or a higher hazard than a lower clock pulse (for example, when the note tone is represented as a rhythmic clock pulse).
- the audio signals of the objects thus generated are then reproduced, for example, by the above-mentioned WFS or ZAP, whereby automatically far distant objects become quieter.
- non-hazardous objects are completely blanked out (not shown) so as not to overload the helmsman or the listener with too much information.
- the playback location may appear at the same distance as the actual distance, ie if the object is one kilometer away according to the radar, the audio object is perceptible at a distance of one kilometer (1: 1 mapping).
- the reproduction location is scaled accordingly so that, for example, a 1: 100 mapping is made and an object one kilometer away is acoustically perceptible or reproduced by an acoustic signal (virtual sound source) approximately ten meters away.
- the former (the 1: 1 figure) has the advantage that no parallax errors occur in the WFS, so that the distance of the object is coded only by the volume and not by the curved waveform.
- Very distant objects would only be audible very late due to the speed of sound, and furthermore, in a 1: 1 representation, very distant objects are hardly distinguishable by distance.
- Exemplary embodiments thus pursue the goal of coding objects with audio signals, so that they can be located as well as possible.
- the audio signals should be sufficiently broadband, since, for example, a sine wave is hardly perceptible.
- narrowband noise or speech should be used to identify objects, not sinusoidal ones.
- pulsed signals are emitted instead of continuous signals (eg a continuous tone).
- the pulse rate can rise similarly to parking sensors in cars with increasing risk.
- the audio signals should sound pleasant when the danger is sufficiently low.
- the danger threshold above which there is a serious danger or below which there is little or no danger potential, is set variably in accordance with the circumstances, for example.
- the danger threshold can optionally also be adapted by the user. For example, the size and speed of a ship or the speeds of the other objects play a role.
- the threshold value can be determined, for example, from the ratio of the time duration to a predicted collision to a braking time of the ship.
- the pleasant sound of the audio signals can be achieved, for example, in that unidentified
- Objects eg objects that pose no danger
- a low center frequency of the narrowband noise or a low pulse rate (rare representation) is used.
- a spectral coloring of the narrowband noise can be used, with the high frequencies less
- the reporting signal may optionally be selected to be precisely located and distinguishable from ambient noise. Moreover, it is advantageous if the reporting signal has a pleasant sound, so that even with long trips, the system is permanently accepted.
- An essential advantage of acoustic, spatially resolving displays is that, unlike optical displays, they can be used by one person simultaneously with the natural environment.
- the natural environment may include, for example, driving on sight or listening to ships and buoys. Thus, a so-called augmented reality can be generated.
- Embodiments are particularly advantageous because they provide an important synergy effect between acoustic and visual display. Namely, the audible indication is always reported and perceived, whereby prioritization for danger may occur while the visual indication requires the attention of the personnel on the bridge. For example, a helmsman sees an object on the radar screen only when he looks at the radar screen. At the same time, however, he no longer looks out of the window and thus loses some of the information about what is happening in his immediate surroundings. Acoustic displays allow him to simultaneously use the information from the radar and the view from the window. Especially in the case of non-self-identifying objects, however, the experienced evaluator is able to classify an object from the radar image (eg as a ship, island or picture disturbance).
- an object from the radar image eg as a ship, island or picture disturbance
- Fig. 1 is a schematic representation of an acoustic display device according to an embodiment of the present invention
- FIG. 2 shows an illustration of a system according to the invention with a sensor for determining the position of an object
- 3a shows representations of location-dependent signals in order to acoustically perceive an increasing danger
- FIG. 4 shows an exemplary embodiment with a multiplicity of loudspeakers for the acoustic representation of two separate objects
- FIG. 5 is a schematic representation of a playback room with a WFS module
- FIG. 6 shows a basic block diagram of a wave field synthesis system with wave field synthesis modules and loudspeaker arrays in a reproduction room.
- FIG. 1 shows a schematic representation of an acoustic display device 100, which has an input 105 above the position information of an object in the Device 100 can be entered.
- the apparatus 100 further has outputs for a plurality of loudspeaker signals LS (for example for a first loudspeaker signal LS1, a second loudspeaker signal LS2, a third loudspeaker signal LS3, ..., an nth loudspeaker signal LSn).
- the input for the position information 105 is designed to signal objects with their position to a signal allocation device 110.
- the signal allocation device 110 is designed to assign an acoustic signal to the objects, wherein the signal allocation device 110 optionally accesses a signal database 140 in order to assign different signals to different objects, for example on the basis of their potential dangers.
- the respectively assigned signal may, for example, depend on whether the object is moving, if so at what speed, or if it is immovable.
- the device 100 has a loudspeaker drive device 120, which receives from the signal allocation device 110 the position of the object and the acoustic signal in order to determine one or more loudspeaker signals LS for a plurality of loudspeakers and these via the outputs for the loudspeaker signals LS1 , ..., LSn output.
- the loudspeaker driver 120 is configured to determine the one or more loudspeaker signals LS based on the acoustic signal assigned to the object. The determination is carried out in such a way that, when the one or more loudspeaker signals LS are reproduced, the position of the object in the reproduction room is indicated acoustically.
- a listener or user then takes the position (eg, distance and direction) of the object as a virtual sound source position.
- one embodiment relates to the reproduction of information of a radar device which determines positions of objects.
- information from, for example, other sources such as sonar or other sensors are implemented in a similar way.
- loudspeakers on the bridge of the ship below windows may be arranged on all walls. These loudspeakers, for example, can all be equipped with their own amplifiers or A / D converters (analog-to-digital converters) and can also be individually controlled.
- FIG. 2 shows a schematic representation of a playback room 210 with three loudspeakers 220a, 220b and 220c and a radar device 230.
- the radar device 230 is connected to the input 105 and provides position information about objects in an environment of the playback room 210.
- the radar device 230 configured to pass the position of the object 200 to the device 100 for acoustic display.
- the three speakers 220a, 220b, 220c are also connected to the outputs for the loudspeaker signals LS of the acoustic display device 100.
- a first speaker 220a is connected to the output for the first speaker signal LS1
- a second speaker 220b is connected to the output for the second speaker signal LS2
- a third speaker 220c is connected to the output for the third speaker signal LS3.
- the acoustic display device 100 evaluates the position information of the object 200 received from the radar 230 to generate three loudspeaker signals LS1, LS2, LS3 for the first, second and third loudspeakers 220a, 220b, 220c. The determination is made such that the position of the object 200 is audible to the listener in the playback room 210, which is at a position P, for example. For this purpose, first the device 100 determines an acoustic signal for the object 200 as a function of the position of the object 200. The position is determined by the distance d and the direction, which can be given for example by an angle ⁇ . Next, the apparatus 100 calculates loudspeaker signals LS for the first to third loudspeakers 220a to 220c.
- This may include, for example, scaling the signal level and delaying the signal so that the listener at position P perceives the object 200 according to its position. For example, in the embodiment shown in FIG. 2, this may occur such that the third loudspeaker 220c provides the strongest signal, during which the first loudspeaker 220a provides only a small signal and the second loudspeaker 220b does not provide a signal.
- the radar device 230 shown in FIG. 2 can also be coupled to a sonar device, which detects, for example, the underwater topography and possibly signals existing shoals that can also be displayed acoustically. To distinguish between different objects (over water, under water or land objects) as mentioned different acoustic signals can be assigned.
- FIGS. 3a and 3b show possible variations of the acoustic signal as a function of the distance of the object and the danger potential associated therewith.
- FIG. 3 a shows a dependence of a frequency f of the signal on the distance d of the object 200.
- a critical distance d c is less than that, there is an increased danger which requires an increased attention of the helmsman.
- This transition from a safe to a dangerous state for example, be signaled in a changing acoustic signal.
- the frequency f of the signal may be close to, or only slightly above, a fundamental frequency f 0 , the frequency range thus defined being perceived as safe by the helmsman.
- the frequency f of the acoustic signal can suddenly rise sharply, so that the increasing danger is signaled to the helmsman.
- the increase in frequency can optionally also increase monotonously with decreasing distance of the object without causing a sudden change in the critical distance and a constantly increasing danger potential for the helmsman becomes perceptible.
- the acoustic signal or the frequency f of the acoustic signal can on the one hand include the audio frequency or else the clock frequency, for example, if the acoustic signal indicates a specific clock in a particular frequency (repetition rate of the clocks). Even with the clock signal, the clock frequency can increase with decreasing distance, so that acoustically an increasing danger potential for the pilot becomes perceptible.
- Fig. 3b shows an embodiment in which the signal level S is shown as a function of time t.
- the distance between two adjacent clocks decreases, so that the clock frequency increases, so that an approaching object will signal.
- the decreasing pitch can be combined by the fact that the signal pulses 63
- the change of the signal may, for example, have a shift of the center frequency to higher frequencies, so that the increasing danger potential is also perceptible in the frequency level or audio frequency of the signal pulses.
- the amplitude or loudness of the signal can increase at the same time as the risk potential increases.
- the acoustic signals are barely perceptible, so that the helmsman is not disturbed by the acoustic signals.
- FIG. 4 shows an embodiment in which a plurality of loudspeakers 220, a first loudspeaker 220a,..., A fourth loudspeaker 22Od,..., A ninth loudspeaker 22Oi,... Have a twelfth loudspeaker 2201.
- the loudspeakers 220 are arranged around the position P of a listener so that the position of an object 200 or the direction of the object 200 becomes noticeable by the fact that only one loudspeaker is active.
- the position of the active loudspeaker corresponds at the same time in the direction of the object 200. This is particularly advantageous when the position P in the reproduction room 210 is fixed.
- a first object 200a at a distance d1 and a second object 200b at a distance d2 from the listening point P may be perceived by the fourth speaker 22Od generating a first sound signal S1 and the second ninth speaker 22Oi generates a second sound signal S2.
- the listener at the position P takes the first object 200a and the second object was then according to their positions.
- the speaker can be selected, which is the shortest distance to the connecting line between the respective object and the Position P has. That would be the fourth speaker 22Od for the first object 200a and the ninth speaker 22Oi for the second object 200b. All other loudspeakers are further away from the respective connection lines (measured as a vertical distance) and, for example, can not be active in this embodiment (do not generate a sound signal).
- the respective adjacent loudspeakers between which the connecting line between the first object 200a and the position P runs, to be active.
- other neighbors speakers may be active. This means that, for example, in further embodiments not only the fourth loudspeaker 22Od is active, but at the same time the third loudspeaker 220c and / or the second loudspeaker 220b and / or the fifth loudspeaker 22Oe can also be active. However, if multiple speakers are simultaneously active to represent the position of one of the objects 200, the amplitude / phase should be selected such that for a listener at position P, the object 200 will be acoustically perceivable at its respective position.
- acoustic perceptibility means that the object 200 is perceived as a virtual sound source, wherein the distance in addition to the volume can also be signaled by a different clock frequency or audio frequency (as was shown, for example, in FIGS. 3a, b).
- FIG. 5 shows an exemplary embodiment in which the loudspeakers are arranged in the context of a wave field synthesis system, so that the acoustic display device 100 drives a first loudspeaker array 221a, a second loudspeaker array 221b and a third loudspeaker array 221c.
- Each of the three loudspeaker arrays 221a, 221b, 221c has, for example, a multiplicity of loudspeakers which, for B.
- each loudspeaker in a respective array can be controlled individually, so that the three arrays, which may be arranged, for example, on the sidewalls of the reproduction room 210, synthesize a wave field which would produce an object 200 as a virtual sound source in the reproduction room 210.
- the device 100 can in turn be coupled to a radar device or a sonar device 230 which transmits the device 100 the position of the respective objects.
- the object itself does not need to be a sound source, but instead a sound signal is specifically assigned to the object. In this sense, therefore, the acoustic display differs according to embodiments of conventional audio playback systems.
- Wave field synthesis is an audio reproduction method developed at TU Delft for the spatial reproduction of complex audio scenes.
- the spatially correct rendering is not limited to a small area, but extends over a wide viewing area.
- WFS is based on a well-founded mathematical-physical basis, namely the principle of Huygens and the Kirchhoff-Helmholtz integral.
- a WFS reproduction system consists of a large number of loudspeakers (so-called secondary sources).
- the loudspeaker signals are formed from delayed and scaled input signals. Since many audio objects (primary sources) are typically used in a WFS scene, many such operations are required to generate the loudspeaker signals. This requires the high computing power required for wave field synthesis.
- WFS also offers the possibility of realistically mapping moving sources. This feature is used in many WFS systems and is included For example, for use in the cinema, virtual reality applications or live performances of great importance.
- a primary goal is the development of signal processing algorithms for the playback of moving sources using WFS.
- the real-time capability of the algorithms is an important condition.
- the most important criterion for evaluating the algorithms is the objective perceived audio quality.
- WFS is a very expensive audio reproduction process in terms of processing resources. This is mainly due to the large number of speakers in a WFS setup and the often high number of virtual sources used in WFS scenes. For this reason, the efficiency of the algorithms to be developed is of paramount importance.
- Wave field synthesis systems have the advantage, in comparison to conventional multi-speaker systems, that exact positioning becomes possible as a result and exact positioning can also be determined at different positions within the reproduction space 210.
- FIG. 6 shows a basic structure of a wave field synthesis system and has a loudspeaker array 221 which is placed relative to a reproduction space 210.
- the loudspeaker array shown in FIG. 6, which is a 360 ° array, includes four array sides 221a, 221b, 221c, and 221d.
- the playback room 210 z. B. a bridge on a ship, it is assumed that with respect to the conventions front / rear or right / left the pre-alignment of the ship is on the same side of the display room 210 where the sub-array 221c is located. In this case, the user who is at the so-called optimal point P in the playback space 210 would see, for example, forward.
- the sub-array 221a would then be behind the user, while the sub-array 221d would be located to the left of the viewer, and the sub-array 221b would be located to the right of the user.
- Each loudspeaker array 221 consists of a number of different individual loudspeakers 708 which are each driven with their own loudspeaker signals LS which are provided by a wave field synthesis module 710 via a data bus 712 shown only schematically in FIG.
- the position information is determined, for example, by a sensor for determining the position of objects (eg the radar) and provided to the wave field synthesis module via the input 105.
- the wave field synthesis module can also receive further inputs, such as, for example, information about the room acoustics of the playback room 210, etc.
- the signal allocator 110 is configured to associate acoustic signals to a plurality of objects 200
- the loudspeaker driver 120 is configured to generate component signals for each of the plurality of objects 200 Combine component signals to speaker signals LS, so that the plurality of objects 200 are acoustically perceptible at different positions.
- the various objects can appear or be perceived as virtual sources (sound sources) for the listeners.
- boundary conditions are considered in the ships.
- the boundary conditions include, for example, requirements for the frequency of the messages, possible positions of the loudspeakers, the required sound pressure level, the characterization of the noise (for example from the engine) and a specification of the control signals for the acoustic display.
- optimal message signals can then be generated taking into account typical spatial sounds on the ships.
- the acoustic drive includes techniques such as binaural coding or the wave field synthesis described above.
- the different techniques are used on test rigs in ships (or one-to-one models of the bridge and / or the control room). For example, psychoacoustic experiments can provide clues.
- Embodiments use reporting signals that are as well as possible to locate in the ship environment, but at the same time sound as pleasant as possible. In this case test setups in the laboratory or else a one-to-one model from the bridge and / or the control station or in vehicles as well as psychoacoustic experiments are useful.
- Further embodiments also provide a connection of sensors and information, for example, from Radar, sounder and nautical charts are received, to the audible indicator.
- An essential part of the connection is the selection of the relevant objects, which should be displayed for example by means of acoustic display.
- embodiments include the following aspects:
- the described systems can also be applied in automobiles, i. Further embodiments also include corresponding driver assistance systems in the car. For example, vehicles approaching laterally (eg when changing lanes) can be signaled acoustically.
- the inventive scheme can also be implemented in software.
- the implementation may be on a digital storage medium, in particular a floppy disk or a CD with electronically readable control signals, which may be provided with a programmable computer system can work together to perform the appropriate procedure.
- the invention thus also consists in a computer program product with program code stored on a machine-readable carrier for carrying out the method according to the invention when the computer program product runs on a computer.
- the invention can thus be realized as a computer program with a program code for carrying out the method when the computer program runs on a computer.
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Acoustics & Sound (AREA)
- Signal Processing (AREA)
- Circuit For Audible Band Transducer (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
- Radar Systems Or Details Thereof (AREA)
Abstract
Description
Claims
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US3820908P | 2008-03-20 | 2008-03-20 | |
PCT/EP2009/001963 WO2009115299A1 (de) | 2008-03-20 | 2009-03-17 | Vorrichtung und verfahren zur akustischen anzeige |
Publications (2)
Publication Number | Publication Date |
---|---|
EP2255359A1 true EP2255359A1 (de) | 2010-12-01 |
EP2255359B1 EP2255359B1 (de) | 2015-07-15 |
Family
ID=40673888
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP09721864.8A Active EP2255359B1 (de) | 2008-03-20 | 2009-03-17 | Vorrichtung und verfahren zur akustischen anzeige |
Country Status (6)
Country | Link |
---|---|
US (1) | US20110188342A1 (de) |
EP (1) | EP2255359B1 (de) |
JP (1) | JP2011516830A (de) |
KR (1) | KR20100116223A (de) |
CN (1) | CN101978424B (de) |
WO (1) | WO2009115299A1 (de) |
Families Citing this family (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101285391B1 (ko) * | 2010-07-28 | 2013-07-10 | 주식회사 팬택 | 음향 객체 정보 융합 장치 및 방법 |
AU2012279357B2 (en) | 2011-07-01 | 2016-01-14 | Dolby Laboratories Licensing Corporation | System and method for adaptive audio signal generation, coding and rendering |
DE102011082310A1 (de) * | 2011-09-07 | 2013-03-07 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung, Verfahren und elektroakustisches System zur Nachhallzeitverlängerung |
KR101308588B1 (ko) * | 2012-02-28 | 2013-09-23 | 주식회사 부국하이텍 | 레이더 시스템 및 레이더 시스템을 이용한 표적의 음파 표시 방법 |
WO2013142657A1 (en) * | 2012-03-23 | 2013-09-26 | Dolby Laboratories Licensing Corporation | System and method of speaker cluster design and rendering |
UA114793C2 (uk) * | 2012-04-20 | 2017-08-10 | Долбі Лабораторіс Лайсензін Корпорейшн | Система та спосіб для генерування, кодування та представлення даних адаптивного звукового сигналу |
EP3317879B1 (de) | 2015-06-30 | 2020-02-19 | Fraunhofer Gesellschaft zur Förderung der Angewand | Verfahren und vorrichtung zum zuordnen von geräuschen und zum analysieren |
GB2542846A (en) * | 2015-10-02 | 2017-04-05 | Ford Global Tech Llc | Hazard indicating system and method |
PT109485A (pt) * | 2016-06-23 | 2017-12-26 | Inst Politécnico De Leiria | Método e aparelho de criação de um cenário tridimensional |
JP7226330B2 (ja) * | 2017-11-01 | 2023-02-21 | ソニーグループ株式会社 | 情報処理装置、情報処理方法及びプログラム |
DE102019006679A1 (de) * | 2019-09-23 | 2021-03-25 | Mbda Deutschland Gmbh | System und Verfahren zur Situationserkennung bezüglich sich in einem Überwachungsraum befindlicher mobiler Objekte |
CN112911354B (zh) * | 2019-12-03 | 2022-11-15 | 海信视像科技股份有限公司 | 显示设备和声音控制方法 |
Family Cites Families (19)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS6299879U (de) * | 1985-12-13 | 1987-06-25 | ||
JP2580117B2 (ja) * | 1985-12-18 | 1997-02-12 | ソニー株式会社 | 再生装置における音場拡大装置 |
JPS6325666U (de) * | 1986-03-13 | 1988-02-19 | ||
FR2744871B1 (fr) * | 1996-02-13 | 1998-03-06 | Sextant Avionique | Systeme de spatialisation sonore, et procede de personnalisation pour sa mise en oeuvre |
US5979586A (en) * | 1997-02-05 | 1999-11-09 | Automotive Systems Laboratory, Inc. | Vehicle collision warning system |
US6097285A (en) * | 1999-03-26 | 2000-08-01 | Lucent Technologies Inc. | Automotive auditory feedback of changing conditions outside the vehicle cabin |
WO2001055833A1 (en) * | 2000-01-28 | 2001-08-02 | Lake Technology Limited | Spatialized audio system for use in a geographical environment |
JP4304845B2 (ja) * | 2000-08-03 | 2009-07-29 | ソニー株式会社 | 音声信号処理方法及び音声信号処理装置 |
DE10155742B4 (de) * | 2001-10-31 | 2004-07-22 | Daimlerchrysler Ag | Vorrichtung und Verfahren zur Generierung von räumlich lokalisierten Warn- und Informationssignalen zur vorbewussten Verarbeitung |
FR2842064B1 (fr) * | 2002-07-02 | 2004-12-03 | Thales Sa | Systeme de spatialisation de sources sonores a performances ameliorees |
US20050222844A1 (en) * | 2004-04-01 | 2005-10-06 | Hideya Kawahara | Method and apparatus for generating spatialized audio from non-three-dimensionally aware applications |
EP1584901A1 (de) * | 2004-04-08 | 2005-10-12 | Wolfgang Dr. Sassin | Vorrichtung zur dynamischen optischen, akustischen oder haptischen Darstellung der Umgebung eines Fahrzeuges |
US8494861B2 (en) * | 2004-05-11 | 2013-07-23 | The Chamberlain Group, Inc. | Movable barrier control system component with audible speech output apparatus and method |
JP2006005868A (ja) * | 2004-06-21 | 2006-01-05 | Denso Corp | 車両用報知音出力装置及びプログラム |
JP2006019908A (ja) * | 2004-06-30 | 2006-01-19 | Denso Corp | 車両用報知音出力装置及びプログラム |
DE102005008333A1 (de) * | 2005-02-23 | 2006-08-31 | Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. | Vorrichtung und Verfahren zum Steuern einer Wellenfeldsynthese-Rendering-Einrichtung |
US20060256976A1 (en) * | 2005-05-11 | 2006-11-16 | House William N | Spatial array monitoring system |
JP4914057B2 (ja) * | 2005-11-28 | 2012-04-11 | 日本無線株式会社 | 船舶用障害物警報装置 |
US7898423B2 (en) * | 2007-07-31 | 2011-03-01 | At&T Intellectual Property I, L.P. | Real-time event notification |
-
2009
- 2009-03-17 CN CN2009801100998A patent/CN101978424B/zh not_active Expired - Fee Related
- 2009-03-17 WO PCT/EP2009/001963 patent/WO2009115299A1/de active Application Filing
- 2009-03-17 JP JP2011500111A patent/JP2011516830A/ja active Pending
- 2009-03-17 KR KR1020107021102A patent/KR20100116223A/ko not_active IP Right Cessation
- 2009-03-17 US US12/922,910 patent/US20110188342A1/en not_active Abandoned
- 2009-03-17 EP EP09721864.8A patent/EP2255359B1/de active Active
Non-Patent Citations (1)
Title |
---|
See references of WO2009115299A1 * |
Also Published As
Publication number | Publication date |
---|---|
CN101978424A (zh) | 2011-02-16 |
KR20100116223A (ko) | 2010-10-29 |
CN101978424B (zh) | 2012-09-05 |
EP2255359B1 (de) | 2015-07-15 |
JP2011516830A (ja) | 2011-05-26 |
US20110188342A1 (en) | 2011-08-04 |
WO2009115299A1 (de) | 2009-09-24 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP2255359B1 (de) | Vorrichtung und verfahren zur akustischen anzeige | |
EP3005732B1 (de) | Vorrichtung und verfahren zur raumselektiven audiowiedergabe | |
DE60217809T2 (de) | Sicherheitsvorrichtung für Fahrzeuge mit einem Mehrkanal-Audio-System | |
DE3413181C3 (de) | ||
EP1878308B1 (de) | Vorrichtung und verfahren zur generierung und bearbeitung von toneffekten in räumlichen tonwiedergabesystemen mittels einer graphischen benutzerschnittstelle | |
DE2910117C2 (de) | Lautsprecherkombination zur Wiedergabe eines zwei- oder mehrkanalig übertragenen Schallereignisses | |
DE102013204798A1 (de) | Umgebungsinformationsmitteilungsvorrichtung | |
DE102012208825A1 (de) | 3d-audiogerät | |
DE102011082886A1 (de) | Stereoklangwiedergabesystem | |
WO2013045374A1 (de) | Verfahren zur rechnergestützten verarbeitung des nahfeldes eines fahrzeugs | |
DE4134130A1 (de) | Einrichtung zum erweitern und steuern von schallfeldern | |
DE102007034029A1 (de) | Verfahren zur Information eines Beobachters über ein im Einsatz befindliches Einsatzfahrzeug und Anordnung dazu | |
EP3044972B1 (de) | Vorrichtung, verfahren und computerprogramm zur dekorrelation von lautsprechersignalen | |
DE102015221361A1 (de) | Verfahren und Vorrichtung zur Fahrerunterstützung | |
DE102009057981B4 (de) | Verfahren zur Steuerung der akustischen Wahrnehmbarkeit eines Fahrzeugs | |
DE19801205A1 (de) | Akustische Warneinrichtung | |
DE102013214239A1 (de) | Warnvorrichtung für ein Fahrzeug, Verfahren und Fahrzeug | |
DE102014217732B4 (de) | Verfahren zum Assistieren eines Fahrers eines Kraftfahrzeugs, Vorrichtung und System | |
DE102016007873A1 (de) | Vorrichtung und Verfahren zur Erzeugung von akustischen Signalen | |
EP2182744B1 (de) | Wiedergabe eines Schallfeldes in einem Zielbeschallungsbereich | |
EP1552720B1 (de) | Verfahren zum vortäuschen einer bewegung mittels einer akustischen wiedergabeeinrichtung und schallwiedergabeanordnung dafür | |
WO2023016924A1 (de) | Verfahren und system zum erzeugen von geräuschen in einem innenraum basierend auf extrahierten und klassifizierten realen geräuschquellen und für spezifische zielgeräusche akustisch transparentes fahrzeug umfassend ein derartiges system | |
DE102015226045A1 (de) | Verfahren und Steuereinheit zur Wiedergabe eines Audiosignals in einem Fahrzeug | |
DE102019113680B3 (de) | Assistenzsystem und Verfahren zur Unterstützung eines Operateurs | |
DE102019004587A1 (de) | Vorrichtung und Verfahren zur akustischen Umgebungspräsentation für ein Fahrzeug |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20101005 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK TR |
|
AX | Request for extension of the european patent |
Extension state: AL BA RS |
|
DAX | Request for extension of the european patent (deleted) | ||
17Q | First examination report despatched |
Effective date: 20140130 |
|
GRAP | Despatch of communication of intention to grant a patent |
Free format text: ORIGINAL CODE: EPIDOSNIGR1 |
|
INTG | Intention to grant announced |
Effective date: 20150202 |
|
GRAS | Grant fee paid |
Free format text: ORIGINAL CODE: EPIDOSNIGR3 |
|
GRAA | (expected) grant |
Free format text: ORIGINAL CODE: 0009210 |
|
AK | Designated contracting states |
Kind code of ref document: B1 Designated state(s): AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO SE SI SK TR |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: EP Ref country code: GB Ref legal event code: FG4D Free format text: NOT ENGLISH |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: FG4D Free format text: LANGUAGE OF EP DOCUMENT: GERMAN |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: REF Ref document number: 737158 Country of ref document: AT Kind code of ref document: T Effective date: 20150815 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R096 Ref document number: 502009011260 Country of ref document: DE |
|
REG | Reference to a national code |
Ref country code: NL Ref legal event code: MP Effective date: 20150715 |
|
REG | Reference to a national code |
Ref country code: LT Ref legal event code: MG4D |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: LV Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: LT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: NO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20151015 Ref country code: FI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: GR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20151016 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: HR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: ES Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: PT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20151116 Ref country code: PL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 8 |
|
REG | Reference to a national code |
Ref country code: DE Ref legal event code: R097 Ref document number: 502009011260 Country of ref document: DE |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CZ Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: IT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: DK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: SK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: EE Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
PLBE | No opposition filed within time limit |
Free format text: ORIGINAL CODE: 0009261 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: NO OPPOSITION FILED WITHIN TIME LIMIT |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: RO Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
26N | No opposition filed |
Effective date: 20160418 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: IS Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: SI Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: BE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20160331 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MC Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: LU Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20160317 |
|
REG | Reference to a national code |
Ref country code: CH Ref legal event code: PL |
|
REG | Reference to a national code |
Ref country code: IE Ref legal event code: MM4A |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: CH Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20160331 Ref country code: IE Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20160317 Ref country code: LI Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20160331 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 9 |
|
REG | Reference to a national code |
Ref country code: AT Ref legal event code: MM01 Ref document number: 737158 Country of ref document: AT Kind code of ref document: T Effective date: 20160317 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: NL Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: AT Free format text: LAPSE BECAUSE OF NON-PAYMENT OF DUE FEES Effective date: 20160317 Ref country code: MT Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
REG | Reference to a national code |
Ref country code: FR Ref legal event code: PLFP Year of fee payment: 10 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: HU Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT; INVALID AB INITIO Effective date: 20090317 Ref country code: CY Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: MK Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 Ref country code: TR Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
PG25 | Lapsed in a contracting state [announced via postgrant information from national office to epo] |
Ref country code: BG Free format text: LAPSE BECAUSE OF FAILURE TO SUBMIT A TRANSLATION OF THE DESCRIPTION OR TO PAY THE FEE WITHIN THE PRESCRIBED TIME-LIMIT Effective date: 20150715 |
|
P01 | Opt-out of the competence of the unified patent court (upc) registered |
Effective date: 20230524 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: DE Payment date: 20240321 Year of fee payment: 16 Ref country code: GB Payment date: 20240322 Year of fee payment: 16 |
|
PGFP | Annual fee paid to national office [announced via postgrant information from national office to epo] |
Ref country code: FR Payment date: 20240319 Year of fee payment: 16 |