US20080037796A1 - 3d audio renderer - Google Patents

3d audio renderer Download PDF

Info

Publication number
US20080037796A1
US20080037796A1 US11/836,137 US83613707A US2008037796A1 US 20080037796 A1 US20080037796 A1 US 20080037796A1 US 83613707 A US83613707 A US 83613707A US 2008037796 A1 US2008037796 A1 US 2008037796A1
Authority
US
United States
Prior art keywords
panning
recited
delay
channel
sound source
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US11/836,137
Other versions
US8488796B2 (en
Inventor
Jean-Marc Jot
Martin Walsh
Adam PHILP
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Creative Technology Ltd
Original Assignee
Creative Technology Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Creative Technology Ltd filed Critical Creative Technology Ltd
Priority to US11/836,137 priority Critical patent/US8488796B2/en
Assigned to CREATIVE TECHNOLOGY LTD reassignment CREATIVE TECHNOLOGY LTD ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: JOT, JEAN MARC, WALSH, MARTIN, PHILP, ADAM R
Publication of US20080037796A1 publication Critical patent/US20080037796A1/en
Application granted granted Critical
Publication of US8488796B2 publication Critical patent/US8488796B2/en
Active legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S1/00Two-channel systems
    • H04S1/002Non-adaptive circuits, e.g. manually adjustable or static, for enhancing the sound image or the spatial distribution
    • H04S1/005For headphones
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S5/00Pseudo-stereo systems, e.g. in which additional channel signals are derived from monophonic signals by means of phase shifting, time delay or reverberation 
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04SSTEREOPHONIC SYSTEMS 
    • H04S2420/00Techniques used stereophonic systems covered by H04S but not provided for in its groups
    • H04S2420/01Enhancing the perception of the sound image or of the spatial distribution using head related transfer functions [HRTF's] or equivalents thereof, e.g. interaural time difference [ITD] or interaural level difference [ILD]

Definitions

  • the present invention relates to signal processing techniques. More particularly, the present invention relates to methods for processing audio signals.
  • Binaural or multi-channel spatialization processing of audio signals typically requires heavy processing costs for increasing the quality of the virtualization experience, especially for accurate 3-D positional audio rendering, for the incorporation of reverberation and reflections, or for rendering spatially extended sources. It is desirable to provide improved binaural and multi-channel spatialization processing algorithms and architectures while minimizing or reducing the associated additional processing costs.
  • a fractional delay implementation is necessary in order to allow for continuous variation of the ITD according to the position of a virtual source.
  • the first-order linear interpolation technique causes significant spectral inaccuracies at high frequencies (a low-pass filtering for non-integer delay values). Avoiding this artifact requires a more expensive fractional delay implementation. It is therefore desirable to provide new techniques for simulating continuous ITD variation that do not require interpolation or fractional delay implementation.
  • Binaural 3D audio simulation is generally based on the synthesis of primary sources that are point source emitters, i.e. which appear to emanate from a single direction in 3D auditory space. In real-world conditions, many sound sources generally approximate the behavior of point sources. However, some sound-emitting objects radiate acoustic energy from a finite surface area or volume whose dimensions render the point-source approximation unacceptable for realistic 3D audio simulation. Such sound-emitting objects may be more suitably represented as line source emitters (such as a vibrating violin string), area source emitters (such as a resonating panel) or volume source emitters (for example a waterfall).
  • line source emitters such as a vibrating violin string
  • area source emitters such as a resonating panel
  • volume source emitters for example a waterfall
  • a spatially extended source In general, the position, shape and dimensions of a spatially extended source are specified and altered under program control, while an appropriate processing algorithm is applied to a monophonic input signal in order to simulate the spatial extent of the emitter.
  • Two existing approaches to this problem include pseudo-stereo approaches and multi-source dynamic decorrelation approaches.
  • pseudo-stereo techniques are to create a pair of decorrelated signals from a monophonic audio input so as to increase the apparent width of the image when played back over two loudspeakers, compared to direct playback of the monophonic input.
  • These techniques can be adapted to simulate spatially extended sources by panning and/or mixing the decorrelated signals.
  • pseudo-stereo algorithms When applied to the 3D audio simulation of spatially extended sources, pseudo-stereo algorithms have three main limitations: they can generate audible artifacts including timbre coloration and phase distortion; they are designed to generate a pair of decorrelated signals, and are not suitable for generating higher numbers of decorrelated versions of the input signal; and they incur substantial per-source computational costs, as each monophonic source is individually processed to generate decorrelated versions prior to mixing or panning.
  • the multi-source dynamic decorrelation approach addresses some of the above limitations.
  • Multiple decorrelated versions of a monophonic input signal are generated using an approach called dynamic decorrelation, which uses a different sparse FIR filter with different delays and coefficients to produce each decorrelated version of the input signal.
  • the delays and coefficients are chosen such that the sum of the decorrelated versions is equal to the original input signal.
  • the resulting decorrelated signals are individually spatialized in 3-D space to cover an area or volume that corresponds to the dimensions of the object being simulated.
  • This technique is less prone to coloration and phase artifacts than prior pseudo-stereo approaches and less restrictive on the number of decorrelated sources that can be generated. Its main limitation is that it incurs substantial per-source computation costs.
  • each resulting signal must then be spatialized individually.
  • the amount of processing necessary to generate a spatially extended sound object is variable, as the number of decorrelated sources generated depends on factors including the spatial extent and shape of the object, as well as the audible angle subtended by the object with respect to the listener, which varies with its orientation and distance. It is desirable to provide new techniques for computationally efficient simulation of spatially extended sound sources.
  • the present invention provides a new method for simulating spatially extended sound sources.
  • simulation of a spatially extended (“volumetric”) sound source may be achieved for a computational cost comparable to that incurred by a normal point source. This is especially advantageous for implementations of this feature on resource-constrained platforms.
  • the invention provides in one embodiment a method for simulating spatially extended sound sources.
  • a first input signal is panned over a plurality of output channels to generate a first multi-channel directionally encoded signal.
  • a second input signal is panned over the plurality of output channels to generate a second multi-channel directionally encoded signal.
  • the first and second multi-channel directionally encoded signals are combined to generate a plurality of loudspeaker output channels.
  • a bank of decorrelation filters are applied on the loudspeaker output channels.
  • the plurality of loudspeakers comprises at least one of real or virtual loudspeakers.
  • the panning comprises deriving an energy scaling factor associated with each of the output channels.
  • the spatially extended source comprises a plurality of notional elementary sources and the energy scaling factor is derived from the summation of contributions of at least one notional elementary source.
  • the notional sources may have discrete panning weights assigned to them and the summation combines the panning weight contributions of the sources.
  • the at least one of the decorrelation filters may comprise any suitable filter including but not limited to one of an all-pass filter, a reverberation filter, a finite impulse response filter, a infinite impulse response filter, and a frequency-domain processing filter.
  • the least a first and a second of the decorrelation filters may, in selected embodiments, have weakly correlated responses.
  • a binaural encoding module for rendering the position of a sound source.
  • the binaural module is configured to generate at least one left signal and one right signal where at least one of these signals is delayed by an integer number of samples, the amount of the delay depending on the position of the sound source.
  • the binaural module is further configured to update the rendered position of the sound source based on transitioning to a new integer delay value triggered by an updated position of the sound source.
  • the rendering a moving sound source includes triggering multiple successive updates of the position of the sound source.
  • at least one of the left signal and the right signal is delayed by reading signal samples first delay tap position in delay memory and transitioning to a new integer delay value is performed by selecting a second delay tap position in delay memory. Further, scaling down the amplitude of the first delay tap to zero and scaling up the amplitude of the second delay tap occurs over a limited transition time.
  • FIG. 1 is a diagram illustrating an overview of a complete spatialization engine, in accordance with one embodiment of the present invention.
  • FIG. 2 is a diagram illustrating a standard multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • FIG. 3 is a diagram illustrating a binaural multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • FIG. 4 is a diagram illustrating a hybrid multi-channel binaural virtualizer for including additional input bus in standard multi-channel format, in accordance with one embodiment of the present invention.
  • FIG. 5 is a diagram illustrating the panning functions of a multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • FIG. 6 is a diagram illustrating a multi-channel decorrelation filter bank, in accordance with one embodiment of the present invention.
  • FIG. 7 is a diagram illustrating a divergence panning scheme in accordance with one embodiment of the present invention.
  • FIG. 8 is a diagram illustrating the implementation of an ITD synthesis module in accordance with one embodiment of the present invention.
  • FIG. 1 is a diagram illustrating an overview of a complete spatialization engine, in accordance with one embodiment of the present invention.
  • FIG. 1 describes a multi-channel spatialization engine.
  • a 3D source signal 102 feeds at least one of the directional encoders 111 a - 111 d .
  • Each of the directional encoders feeds one of the multi-channel master buses 106 .
  • the directional encoder 11 a feeds a diffuse multichannel mixing bus which feeds a multi-channel decorrelation filter bank 122 .
  • the output of the multi-channel decorrelation filter bank 122 may be fed directly to an array of loudspeaker outputs, or, indirectly, as illustrated in FIG. 1 , to a virtualizer 120 for binaural reproduction over headphones.
  • FIG. 2 describes two 3D source signals 202 and 204 .
  • Each 3D source signal is processed by a directional encoder ( 208 and 210 ).
  • Each directional encoder pans an input signal over a plurality of output channels to generate a first multi-channel directionally encoded signal.
  • the multichannel directionally encoded signals are combined additively into a master bus 212 which directly feeds an array of loudspeaker outputs.
  • Each directional encoder ( 208 ) performs a panning operation by scaling the input signal using amplitude scalers denoted gi. The values of the scalers gi are determined by the desired panning direction ⁇ .
  • FIG. 3 is a diagram illustrating a binaural multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • a 3D source signal 302 is fed to a delay line where it is split into a left signal and a right signal.
  • Each of the left signal and the right signal feeds a multi-channel directional encoder to generate a left multichannel directionally encoded signals and a right multichannel directionally encoded signal into a multi-channel binaural mixing bus 306 .
  • the multi-channel binaural mixing bus feeds a reconstruction filter bank where the individual channel signals are filtered by a set of HRTF filters 308 and combined to produce a left output channel 320 and a right output channel 322 .
  • FIG. 4 is a diagram illustrating a hybrid multi-channel binaural virtualizer 400 corresponding generally to the virtualizer 120 illustrated in FIG. 1 , in accordance with one embodiment of the present invention.
  • the virtualizer 400 processes the left and right multichannel mixing bus signals 402 and 404 in a manner similar to the virtualizer 332 .
  • it receives the standard multi-channel mixing bus 406 , and feeds them to the set of HRTF filters 410 after inserting delays 408 to synthesize the interchannel delays corresponding to each of the virtual loudspeaker positions.
  • FIG. 5 is a diagram illustrating the panning functions of a multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • FIG. 5 shows an example of discrete multi-channel horizontal-only amplitude-preserving panning functions obtained by the VBAP method for the principal direction azimuths ⁇ 0, ⁇ 30, ⁇ 60, ⁇ 90, ⁇ 120, 180 degrees ⁇ .
  • FIG. 6 is a diagram illustrating a multi-channel decorrelation filter bank, in accordance with one embodiment of the present invention.
  • the multi-channel filter bank 604 corresponds generally to block 122 illustrated in FIG. 1 .
  • the multi-channel ‘diffuse’ master bus feeds a multi-channel decorrelation filter bank (such that each channel of the bus feeds a different filter from the bank) while divergence panning is applied on a per-source basis for each spatially extended source.
  • the output of the decorrelation filter bank is mixed into the standard multi-channel bus before virtualization.
  • input signals are received over the diffuse multi-channel bus 602 and filtered by filters 606 - 609 to decorrelate them.
  • the decorrelated output signals 612 are then fed into the standard multi-channel bus 106 illustrated in FIG. 1 .
  • FIG. 7 is a diagram illustrating a divergence panning scheme in accordance with one embodiment of the present invention.
  • the proposed spatialization engine employs a particular type of directional panning algorithm to control the spatial distribution of reverberation components and clustered reflections.
  • this type of algorithm referred to as ‘divergence panning’, controls the angular extent of a radiating arc centered around this direction. This is illustrated in FIG. 7 for the 2-D case.
  • the value of the divergence angle ⁇ div can vary from 0 (pinpoint localization) to ⁇ (diffuse localization).
  • a convenient alternative consists of representing the direction angle and the divergence angle together in the form of a panning vector whose magnitude is 1.0 for pinpoint localization and 0.0 for diffuse localization.
  • the practical implementation of the divergence panning algorithm illustrated in FIG. 7 requires a method for deriving an energy scaling factor associated with each of the output channels. This can be achieved by modeling the radiating arc as a uniform distribution of notional sources with a total energy of 1.0, assigning discrete energy panning weights to each of these notional sources and summing the panning weight contributions of all these sources to derive the desired energy scaling factor for this channel. This method can be readily extended to three dimensions (e.g. by considering an axis-symmetric distribution of sources around the point located at direction ( ⁇ , ⁇ ) on the 3-D sphere).
  • a new method for simulating spatially extended sound sources is provided.
  • This allows simulating a spatially extended (“volumetric”) sound source for a computational cost comparable to that incurred by a normal (point) source. This will be valuable for any implementation of this feature on resource constrained platforms.
  • the only known alternative solutions uses typically 2 or 3 point sources to simulate a volumetric source and requires a per-source dynamic decorrelation algorithm which does not map well to some current audio processors.
  • a multi-channel ‘diffuse’ master bus feeds a multi-channel decorrelation filter bank (such that each channel of the bus feeds a different filter from the bank) while divergence panning is applied on a per-source basis for each spatially extended source, using a directional encoder as illustrated in FIG. 2 (block 208 ), where the scaling factors are computed to realize divergence panning.
  • the output of the decorrelation filter bank is mixed into the standard multi-channel bus before virtualization.
  • This new technique offers several advantages over existing spatially extended source simulation techniques: (1) the per-source processing cost for a spatially extended source is significantly reduced, becoming comparable to that of a point source spatialized in multi-channel binaural mode; (2) the desired spatial extent (divergence angle) can be reproduced precisely regardless of the shape of the object to be simulated; and (3) since the decorrelation filter bank is common to all sources, its cost is not critical and it can be designed without compromises. Ideally, it consists of mutually orthogonal all-pass filters. Alternatively, it can be based on synthetic quasi-colorless reverberation responses.
  • FIG. 8 is a diagram illustrating the implementation of an ITD synthesis module in accordance with one embodiment of the present invention.
  • ITD interaural time delay
  • the transition can be implemented by simple cross-fading between two delay taps or by a time-varying delay implementation using first order linear interpolation.
  • fractional delay filter techniques i.e., fractional delay implementation using FIR interpolator or variable all-pass filter.

Abstract

A method for simulating spatially extended sound sources comprising: panning a first input signal over a plurality of output channels to generate a first multi-channel directionally encoded signal; panning a second input signal over the plurality of output channels to generate a second multi-channel directionally encoded signal; combining the first and second multi-channel directionally encoded signals to generate a plurality of loudspeaker output channels; and applying a bank of decorrelation filters on the loudspeaker output channels.

Description

    CROSS-REFERENCES TO RELATED APPLICATIONS
  • This application claims priority from provisional U.S. Patent Application Ser. No. 60/821,815, filed Aug. 8, 2006, titled “3D Audio Renderer” the disclosure of which is incorporated by reference in its entirety.
  • BACKGROUND OF THE INVENTION
  • 1. Field of the Invention
  • The present invention relates to signal processing techniques. More particularly, the present invention relates to methods for processing audio signals.
  • 2. Description of the Related Art
  • Binaural or multi-channel spatialization processing of audio signals typically requires heavy processing costs for increasing the quality of the virtualization experience, especially for accurate 3-D positional audio rendering, for the incorporation of reverberation and reflections, or for rendering spatially extended sources. It is desirable to provide improved binaural and multi-channel spatialization processing algorithms and architectures while minimizing or reducing the associated additional processing costs.
  • In binaural 3-D positional audio rendering schemes, a fractional delay implementation is necessary in order to allow for continuous variation of the ITD according to the position of a virtual source. The first-order linear interpolation technique causes significant spectral inaccuracies at high frequencies (a low-pass filtering for non-integer delay values). Avoiding this artifact requires a more expensive fractional delay implementation. It is therefore desirable to provide new techniques for simulating continuous ITD variation that do not require interpolation or fractional delay implementation.
  • Binaural 3D audio simulation is generally based on the synthesis of primary sources that are point source emitters, i.e. which appear to emanate from a single direction in 3D auditory space. In real-world conditions, many sound sources generally approximate the behavior of point sources. However, some sound-emitting objects radiate acoustic energy from a finite surface area or volume whose dimensions render the point-source approximation unacceptable for realistic 3D audio simulation. Such sound-emitting objects may be more suitably represented as line source emitters (such as a vibrating violin string), area source emitters (such as a resonating panel) or volume source emitters (for example a waterfall).
  • In general, the position, shape and dimensions of a spatially extended source are specified and altered under program control, while an appropriate processing algorithm is applied to a monophonic input signal in order to simulate the spatial extent of the emitter. Two existing approaches to this problem include pseudo-stereo approaches and multi-source dynamic decorrelation approaches.
  • The goal of pseudo-stereo techniques is to create a pair of decorrelated signals from a monophonic audio input so as to increase the apparent width of the image when played back over two loudspeakers, compared to direct playback of the monophonic input. These techniques can be adapted to simulate spatially extended sources by panning and/or mixing the decorrelated signals. When applied to the 3D audio simulation of spatially extended sources, pseudo-stereo algorithms have three main limitations: they can generate audible artifacts including timbre coloration and phase distortion; they are designed to generate a pair of decorrelated signals, and are not suitable for generating higher numbers of decorrelated versions of the input signal; and they incur substantial per-source computational costs, as each monophonic source is individually processed to generate decorrelated versions prior to mixing or panning.
  • The multi-source dynamic decorrelation approach addresses some of the above limitations. Multiple decorrelated versions of a monophonic input signal are generated using an approach called dynamic decorrelation, which uses a different sparse FIR filter with different delays and coefficients to produce each decorrelated version of the input signal. The delays and coefficients are chosen such that the sum of the decorrelated versions is equal to the original input signal. The resulting decorrelated signals are individually spatialized in 3-D space to cover an area or volume that corresponds to the dimensions of the object being simulated. This technique is less prone to coloration and phase artifacts than prior pseudo-stereo approaches and less restrictive on the number of decorrelated sources that can be generated. Its main limitation is that it incurs substantial per-source computation costs. Not only must multiple decorrelated signals be generated for each object, but each resulting signal must then be spatialized individually. The amount of processing necessary to generate a spatially extended sound object is variable, as the number of decorrelated sources generated depends on factors including the spatial extent and shape of the object, as well as the audible angle subtended by the object with respect to the listener, which varies with its orientation and distance. It is desirable to provide new techniques for computationally efficient simulation of spatially extended sound sources.
  • SUMMARY OF THE INVENTION
  • The present invention provides a new method for simulating spatially extended sound sources. By using the techniques described herein, simulation of a spatially extended (“volumetric”) sound source may be achieved for a computational cost comparable to that incurred by a normal point source. This is especially advantageous for implementations of this feature on resource-constrained platforms.
  • The invention provides in one embodiment a method for simulating spatially extended sound sources. A first input signal is panned over a plurality of output channels to generate a first multi-channel directionally encoded signal. A second input signal is panned over the plurality of output channels to generate a second multi-channel directionally encoded signal. The first and second multi-channel directionally encoded signals are combined to generate a plurality of loudspeaker output channels. A bank of decorrelation filters are applied on the loudspeaker output channels.
  • In accordance with variations of this embodiment, the plurality of loudspeakers comprises at least one of real or virtual loudspeakers. In accordance with another embodiment, the panning comprises deriving an energy scaling factor associated with each of the output channels. The spatially extended source comprises a plurality of notional elementary sources and the energy scaling factor is derived from the summation of contributions of at least one notional elementary source. The notional sources may have discrete panning weights assigned to them and the summation combines the panning weight contributions of the sources. In yet other embodiments, the at least one of the decorrelation filters may comprise any suitable filter including but not limited to one of an all-pass filter, a reverberation filter, a finite impulse response filter, a infinite impulse response filter, and a frequency-domain processing filter. The least a first and a second of the decorrelation filters may, in selected embodiments, have weakly correlated responses.
  • In accordance with another embodiment, a binaural encoding module for rendering the position of a sound source is provided. The binaural module is configured to generate at least one left signal and one right signal where at least one of these signals is delayed by an integer number of samples, the amount of the delay depending on the position of the sound source. The binaural module is further configured to update the rendered position of the sound source based on transitioning to a new integer delay value triggered by an updated position of the sound source.
  • In accordance with another embodiment, the rendering a moving sound source includes triggering multiple successive updates of the position of the sound source. In accordance with yet another embodiment at least one of the left signal and the right signal is delayed by reading signal samples first delay tap position in delay memory and transitioning to a new integer delay value is performed by selecting a second delay tap position in delay memory. Further, scaling down the amplitude of the first delay tap to zero and scaling up the amplitude of the second delay tap occurs over a limited transition time.
  • These and other features and advantages of the present invention are described below with reference to the drawings.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is a diagram illustrating an overview of a complete spatialization engine, in accordance with one embodiment of the present invention.
  • FIG. 2 is a diagram illustrating a standard multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • FIG. 3 is a diagram illustrating a binaural multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • FIG. 4 is a diagram illustrating a hybrid multi-channel binaural virtualizer for including additional input bus in standard multi-channel format, in accordance with one embodiment of the present invention.
  • FIG. 5 is a diagram illustrating the panning functions of a multi-channel directional encoder, in accordance with one embodiment of the present invention.
  • FIG. 6 is a diagram illustrating a multi-channel decorrelation filter bank, in accordance with one embodiment of the present invention.
  • FIG. 7 is a diagram illustrating a divergence panning scheme in accordance with one embodiment of the present invention.
  • FIG. 8 is a diagram illustrating the implementation of an ITD synthesis module in accordance with one embodiment of the present invention.
  • DETAILED DESCRIPTION OF PREFERRED EMBODIMENTS
  • Reference will now be made in detail to preferred embodiments of the invention. Examples of the preferred embodiments are illustrated in the accompanying drawings. While the invention will be described in conjunction with these preferred embodiments, it will be understood that it is not intended to limit the invention to such preferred embodiments. On the contrary, it is intended to cover alternatives, modifications, and equivalents as may be included within the spirit and scope of the invention as defined by the appended claims. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. The present invention may be practiced without some or all of these specific details. In other instances, well known mechanisms have not been described in detail in order not to unnecessarily obscure the present invention.
  • It should be noted herein that throughout the various drawings like numerals refer to like parts. The various drawings illustrated and described herein are used to illustrate various features of the invention. To the extent that a particular feature is illustrated in one drawing and not another, except where otherwise indicated or where the structure inherently prohibits incorporation of the feature, it is to be understood that those features may be adapted to be included in the embodiments represented in the other figures, as if they were fully illustrated in those figures. Unless otherwise indicated, the drawings are not necessarily to scale. Any dimensions provided on the drawings are not intended to be limiting as to the scope of the invention but merely illustrative.
  • FIG. 1 is a diagram illustrating an overview of a complete spatialization engine, in accordance with one embodiment of the present invention. FIG. 1 describes a multi-channel spatialization engine. A 3D source signal 102 feeds at least one of the directional encoders 111 a-111 d. Each of the directional encoders feeds one of the multi-channel master buses 106. The directional encoder 11 a feeds a diffuse multichannel mixing bus which feeds a multi-channel decorrelation filter bank 122. The output of the multi-channel decorrelation filter bank 122 may be fed directly to an array of loudspeaker outputs, or, indirectly, as illustrated in FIG. 1, to a virtualizer 120 for binaural reproduction over headphones.
  • FIG. 2 describes two 3D source signals 202 and 204. Each 3D source signal is processed by a directional encoder (208 and 210). Each directional encoder pans an input signal over a plurality of output channels to generate a first multi-channel directionally encoded signal. The multichannel directionally encoded signals are combined additively into a master bus 212 which directly feeds an array of loudspeaker outputs. Each directional encoder (208) performs a panning operation by scaling the input signal using amplitude scalers denoted gi. The values of the scalers gi are determined by the desired panning direction θ.
  • FIG. 3 is a diagram illustrating a binaural multi-channel directional encoder, in accordance with one embodiment of the present invention. A 3D source signal 302 is fed to a delay line where it is split into a left signal and a right signal. Each of the left signal and the right signal feeds a multi-channel directional encoder to generate a left multichannel directionally encoded signals and a right multichannel directionally encoded signal into a multi-channel binaural mixing bus 306. The multi-channel binaural mixing bus feeds a reconstruction filter bank where the individual channel signals are filtered by a set of HRTF filters 308 and combined to produce a left output channel 320 and a right output channel 322.
  • FIG. 4 is a diagram illustrating a hybrid multi-channel binaural virtualizer 400 corresponding generally to the virtualizer 120 illustrated in FIG. 1, in accordance with one embodiment of the present invention. The virtualizer 400 processes the left and right multichannel mixing bus signals 402 and 404 in a manner similar to the virtualizer 332. In addition, it receives the standard multi-channel mixing bus 406, and feeds them to the set of HRTF filters 410 after inserting delays 408 to synthesize the interchannel delays corresponding to each of the virtual loudspeaker positions.
  • FIG. 5 is a diagram illustrating the panning functions of a multi-channel directional encoder, in accordance with one embodiment of the present invention. The set of N-channel spatial panning functions {gi(*, *), i=0, 1, . . . N−1} is considered ‘discrete’ if, for any direction (*, *), there are at most three non-zero panning functions and if, for each panning function gi, there is a ‘principal direction’ (*i, *i) where this panning function reaches its maximum value and is the only non-zero panning function in the set. Discrete panning functions are computationally advantageous because they minimize the number of non-zero panning weights necessary to synthesize any given direction with the directional encoder of FIG. 2 or FIG. 3. FIG. 5 shows an example of discrete multi-channel horizontal-only amplitude-preserving panning functions obtained by the VBAP method for the principal direction azimuths {0, ±30, ±60, ±90, ±120, 180 degrees}.
  • FIG. 6 is a diagram illustrating a multi-channel decorrelation filter bank, in accordance with one embodiment of the present invention. The multi-channel filter bank 604 corresponds generally to block 122 illustrated in FIG. 1. The multi-channel ‘diffuse’ master bus feeds a multi-channel decorrelation filter bank (such that each channel of the bus feeds a different filter from the bank) while divergence panning is applied on a per-source basis for each spatially extended source. The output of the decorrelation filter bank is mixed into the standard multi-channel bus before virtualization. As illustrated, input signals are received over the diffuse multi-channel bus 602 and filtered by filters 606-609 to decorrelate them. The decorrelated output signals 612 are then fed into the standard multi-channel bus 106 illustrated in FIG. 1.
  • Divergence Panning
  • FIG. 7 is a diagram illustrating a divergence panning scheme in accordance with one embodiment of the present invention. The proposed spatialization engine employs a particular type of directional panning algorithm to control the spatial distribution of reverberation components and clustered reflections. In addition to reproducing a direction, this type of algorithm, referred to as ‘divergence panning’, controls the angular extent of a radiating arc centered around this direction. This is illustrated in FIG. 7 for the 2-D case. According to one embodiment, the value of the divergence angle θ div can vary from 0 (pinpoint localization) to π (diffuse localization).
  • A convenient alternative consists of representing the direction angle and the divergence angle together in the form of a panning vector whose magnitude is 1.0 for pinpoint localization and 0.0 for diffuse localization. This property is obtained if the panning vector, denoted s, is defined as the normalized integrated energy vector for a continuous distribution of sound sources on the radiating arc shown in FIG. 1, according to the formalism proposed by Gerzon:
    s∥=[∫ [−θ div,θ div]cos(θ)dθ]/[∫ [−θ div,θ div] dθ].
  • This yields the relation between the panning vector magnitude and the divergence angle θ div in 2D:
    s∥=sin(θ div)/θ div.
  • The practical implementation of the divergence panning algorithm illustrated in FIG. 7 requires a method for deriving an energy scaling factor associated with each of the output channels. This can be achieved by modeling the radiating arc as a uniform distribution of notional sources with a total energy of 1.0, assigning discrete energy panning weights to each of these notional sources and summing the panning weight contributions of all these sources to derive the desired energy scaling factor for this channel. This method can be readily extended to three dimensions (e.g. by considering an axis-symmetric distribution of sources around the point located at direction (θ, φ) on the 3-D sphere).
  • Spatially Extended Sources
  • In accordance with an embodiment of the present invention, a new method for simulating spatially extended sound sources is provided. This allows simulating a spatially extended (“volumetric”) sound source for a computational cost comparable to that incurred by a normal (point) source. This will be valuable for any implementation of this feature on resource constrained platforms. The only known alternative solutions uses typically 2 or 3 point sources to simulate a volumetric source and requires a per-source dynamic decorrelation algorithm which does not map well to some current audio processors.
  • In the architecture of FIG. 1, a multi-channel ‘diffuse’ master bus feeds a multi-channel decorrelation filter bank (such that each channel of the bus feeds a different filter from the bank) while divergence panning is applied on a per-source basis for each spatially extended source, using a directional encoder as illustrated in FIG. 2 (block 208), where the scaling factors are computed to realize divergence panning. The output of the decorrelation filter bank is mixed into the standard multi-channel bus before virtualization.
  • This new technique offers several advantages over existing spatially extended source simulation techniques: (1) the per-source processing cost for a spatially extended source is significantly reduced, becoming comparable to that of a point source spatialized in multi-channel binaural mode; (2) the desired spatial extent (divergence angle) can be reproduced precisely regardless of the shape of the object to be simulated; and (3) since the decorrelation filter bank is common to all sources, its cost is not critical and it can be designed without compromises. Ideally, it consists of mutually orthogonal all-pass filters. Alternatively, it can be based on synthetic quasi-colorless reverberation responses.
  • ITD Synthesis
  • FIG. 8 is a diagram illustrating the implementation of an ITD synthesis module in accordance with one embodiment of the present invention.
  • A computationally efficient method for synthesizing interaural time delay (ITD) cues is provided. This method allows the implementation of a time-varying ITD with no audible artifacts and without using costly fractional delay filter techniques. A computationally efficient ITD implementation is obtained by recognizing that:
  • (1) The simulation of a static arbitrary direction will be satisfactory even if the ITD value is rounded to the nearest integer number of samples, provided that the sample rate be sufficiently high. At a sample rate of 48 kHz, for instance, a difference of 0.5 sample on the ITD (the worst-case rounding error) corresponds approximately to an azimuth difference of 1.5 degrees, which is considered imperceptible.
  • (2) When the position of the virtual source needs to be updated, spectral inaccuracies occurring during the transition to a new position will not be noticeable if this transition is of short enough duration. Therefore, the transition can be implemented by simple cross-fading between two delay taps or by a time-varying delay implementation using first order linear interpolation.
  • Conventional technology would also incur significant additional processing cost per source due to costly fractional delay filter techniques, i.e., fractional delay implementation using FIR interpolator or variable all-pass filter).
  • In practice, it is simpler to introduce the ITD on the contra-lateral path only, leaving the ipsi-lateral path un-delayed. Individual adaptation of the ITD according to the morphology of the listener may be achieved approximately by adjusting the value of the spherical head radius r in Equation (8) or via a more elaborate model.
  • Although the foregoing invention has been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications may be practiced within the scope of the appended claims. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the invention is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.

Claims (11)

1. A method for simulating spatially extended sound sources comprising:
panning a first input signal over a plurality of output channels to generate a first multi-channel directionally encoded signal;
panning a second input signal over the plurality of output channels to generate a second multi-channel directionally encoded signal;
combining the first and second multi-channel directionally encoded signals to generate a plurality of loudspeaker output channels; and
applying a bank of decorrelation filters on the loudspeaker output channels.
2. The method as recited in claim 1 wherein the plurality of loudspeakers comprises at least one of real or virtual loudspeakers.
3. The method as recited in claim 1 wherein the panning comprises deriving an energy scaling factor associated with each of the output channels.
4. The method as recited in claim 3 wherein the spatially extended source comprises a plurality of notional elementary sources and the energy scaling factor is derived from the summation of contributions of at least one notional elementary source.
5. The method as recited in claim 4 wherein the notional sources having discrete panning weights assigned to them and the summation combines the panning weight contributions of the sources.
6. The method as recited in claim 1 wherein at least one of the decorrelation filters is one of an all-pass filter, a reverberation filter, a finite impulse response filter, a infinite impulse response filter, and a frequency-domain processing filter.
7. The method as recited in claim 1 wherein at least a first and a second of the decorrelation filters have weakly correlated responses.
8. The method as recited in claim 1 wherein a spatially extended sound source is represented as a combination of a direction and a divergence angle.
9. A binaural encoding module for rendering the position of a sound source configured to perform the steps of:
generating at least one left signal and one right signal where at least one of these signals is delayed by an integer number of samples, the amount of the delay depending on the position of the sound source; and
updating the position of the sound source includes transitioning to a new integer delay value triggered by an updated position of the sound source.
10. The method as recited in claim 9 wherein rendering a moving sound source includes triggering multiple successive updates of the position of the sound source.
11. The method as recited in claim 9 wherein at least one of the left signal and the right signal is delayed by reading signal samples first delay tap position in delay memory and transitioning to a new integer delay value is performed by selecting a second delay tap position in delay memory
scaling down the amplitude of the first delay tap to zero and scaling up the amplitude of the second delay tap over a limited transition time.
US11/836,137 2006-08-08 2007-08-08 3D audio renderer Active 2031-11-11 US8488796B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/836,137 US8488796B2 (en) 2006-08-08 2007-08-08 3D audio renderer

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US82181506P 2006-08-08 2006-08-08
US11/836,137 US8488796B2 (en) 2006-08-08 2007-08-08 3D audio renderer

Publications (2)

Publication Number Publication Date
US20080037796A1 true US20080037796A1 (en) 2008-02-14
US8488796B2 US8488796B2 (en) 2013-07-16

Family

ID=39050823

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/836,137 Active 2031-11-11 US8488796B2 (en) 2006-08-08 2007-08-08 3D audio renderer

Country Status (1)

Country Link
US (1) US8488796B2 (en)

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20080126104A1 (en) * 2004-08-25 2008-05-29 Dolby Laboratories Licensing Corporation Multichannel Decorrelation In Spatial Audio Coding
US20080247556A1 (en) * 2007-02-21 2008-10-09 Wolfgang Hess Objective quantification of auditory source width of a loudspeakers-room system
WO2010012478A3 (en) * 2008-07-31 2010-04-08 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Signal generation for binaural signals
US20100303246A1 (en) * 2009-06-01 2010-12-02 Dts, Inc. Virtual audio processing for loudspeaker or headphone playback
US20120002024A1 (en) * 2010-06-08 2012-01-05 Lg Electronics Inc. Image display apparatus and method for operating the same
AU2013263871B2 (en) * 2008-07-31 2015-07-30 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Signal generation for binaural signals
US9197979B2 (en) 2012-05-31 2015-11-24 Dts Llc Object-based audio system using vector base amplitude panning
US20160034248A1 (en) * 2014-07-29 2016-02-04 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for conducting interactive sound propagation and rendering for a plurality of sound sources in a virtual environment scene
US9264838B2 (en) 2012-12-27 2016-02-16 Dts, Inc. System and method for variable decorrelation of audio signals
EP2991382A1 (en) * 2014-08-29 2016-03-02 Huawei Technologies Co., Ltd. Sound signal processing method and apparatus
GB2565747A (en) * 2017-04-20 2019-02-27 Nokia Technologies Oy Enhancing loudspeaker playback using a spatial extent processed audio signal
US10248744B2 (en) 2017-02-16 2019-04-02 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for acoustic classification and optimization for multi-modal rendering of real-world scenes
US10679407B2 (en) 2014-06-27 2020-06-09 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for modeling interactive diffuse reflections and higher-order diffraction in virtual environment scenes
US11304020B2 (en) * 2016-05-06 2022-04-12 Dts, Inc. Immersive audio reproduction systems
GB2609667A (en) * 2021-08-13 2023-02-15 British Broadcasting Corp Audio rendering

Families Citing this family (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8908874B2 (en) 2010-09-08 2014-12-09 Dts, Inc. Spatial audio encoding and reproduction
WO2013122387A1 (en) 2012-02-15 2013-08-22 Samsung Electronics Co., Ltd. Data transmitting apparatus, data receiving apparatus, data transceiving system, data transmitting method, and data receiving method
WO2013122385A1 (en) * 2012-02-15 2013-08-22 Samsung Electronics Co., Ltd. Data transmitting apparatus, data receiving apparatus, data transreceiving system, data transmitting method, data receiving method and data transreceiving method
WO2013122386A1 (en) 2012-02-15 2013-08-22 Samsung Electronics Co., Ltd. Data transmitting apparatus, data receiving apparatus, data transreceiving system, data transmitting method, data receiving method and data transreceiving method
CN103901400B (en) * 2014-04-10 2016-08-17 北京大学深圳研究生院 A kind of based on delay compensation and ears conforming binaural sound source of sound localization method
US10325610B2 (en) 2016-03-30 2019-06-18 Microsoft Technology Licensing, Llc Adaptive audio rendering
CN115175064A (en) 2017-10-17 2022-10-11 奇跃公司 Mixed reality spatial audio
JP2021514081A (en) 2018-02-15 2021-06-03 マジック リープ, インコーポレイテッドMagic Leap,Inc. Mixed reality virtual echo
CN112236940A (en) 2018-05-30 2021-01-15 奇跃公司 Indexing scheme for filter parameters
US11070933B1 (en) * 2019-08-06 2021-07-20 Apple Inc. Real-time acoustic simulation of edge diffraction
EP4049466A4 (en) 2019-10-25 2022-12-28 Magic Leap, Inc. Reverberation fingerprint estimation

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5491754A (en) * 1992-03-03 1996-02-13 France Telecom Method and system for artificial spatialisation of digital audio signals
US6011851A (en) * 1997-06-23 2000-01-04 Cisco Technology, Inc. Spatial audio processing method and apparatus for context switching between telephony applications
US6035045A (en) * 1996-10-22 2000-03-07 Kabushiki Kaisha Kawai Gakki Seisakusho Sound image localization method and apparatus, delay amount control apparatus, and sound image control apparatus with using delay amount control apparatus
US6078669A (en) * 1997-07-14 2000-06-20 Euphonics, Incorporated Audio spatial localization apparatus and methods
US6111958A (en) * 1997-03-21 2000-08-29 Euphonics, Incorporated Audio spatial enhancement apparatus and methods
US6498857B1 (en) * 1998-06-20 2002-12-24 Central Research Laboratories Limited Method of synthesizing an audio signal
US6507658B1 (en) * 1999-01-27 2003-01-14 Kind Of Loud Technologies, Llc Surround sound panner
US6714652B1 (en) * 1999-07-09 2004-03-30 Creative Technology, Ltd. Dynamic decorrelator for audio signals
US20060165184A1 (en) * 2004-11-02 2006-07-27 Heiko Purnhagen Audio coding using de-correlated signals
US7174229B1 (en) * 1998-11-13 2007-02-06 Agere Systems Inc. Method and apparatus for processing interaural time delay in 3D digital audio
US7231054B1 (en) * 1999-09-24 2007-06-12 Creative Technology Ltd Method and apparatus for three-dimensional audio display
US7356465B2 (en) * 2003-11-26 2008-04-08 Inria Institut National De Recherche En Informatique Et En Automatique Perfected device and method for the spatialization of sound
US7412380B1 (en) * 2003-12-17 2008-08-12 Creative Technology Ltd. Ambience extraction and modification for enhancement and upmix of audio signals

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5491754A (en) * 1992-03-03 1996-02-13 France Telecom Method and system for artificial spatialisation of digital audio signals
US6035045A (en) * 1996-10-22 2000-03-07 Kabushiki Kaisha Kawai Gakki Seisakusho Sound image localization method and apparatus, delay amount control apparatus, and sound image control apparatus with using delay amount control apparatus
US6111958A (en) * 1997-03-21 2000-08-29 Euphonics, Incorporated Audio spatial enhancement apparatus and methods
US6011851A (en) * 1997-06-23 2000-01-04 Cisco Technology, Inc. Spatial audio processing method and apparatus for context switching between telephony applications
US6078669A (en) * 1997-07-14 2000-06-20 Euphonics, Incorporated Audio spatial localization apparatus and methods
US6498857B1 (en) * 1998-06-20 2002-12-24 Central Research Laboratories Limited Method of synthesizing an audio signal
US7174229B1 (en) * 1998-11-13 2007-02-06 Agere Systems Inc. Method and apparatus for processing interaural time delay in 3D digital audio
US6507658B1 (en) * 1999-01-27 2003-01-14 Kind Of Loud Technologies, Llc Surround sound panner
US6714652B1 (en) * 1999-07-09 2004-03-30 Creative Technology, Ltd. Dynamic decorrelator for audio signals
US7231054B1 (en) * 1999-09-24 2007-06-12 Creative Technology Ltd Method and apparatus for three-dimensional audio display
US7356465B2 (en) * 2003-11-26 2008-04-08 Inria Institut National De Recherche En Informatique Et En Automatique Perfected device and method for the spatialization of sound
US7412380B1 (en) * 2003-12-17 2008-08-12 Creative Technology Ltd. Ambience extraction and modification for enhancement and upmix of audio signals
US20060165184A1 (en) * 2004-11-02 2006-07-27 Heiko Purnhagen Audio coding using de-correlated signals

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
Chandra et al; A Binaural Synthesis with Multiple Sound Sources Based on Spatial Features of Head-related Transfer Functions, IEEE, 2006 *

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8015018B2 (en) * 2004-08-25 2011-09-06 Dolby Laboratories Licensing Corporation Multichannel decorrelation in spatial audio coding
US20080126104A1 (en) * 2004-08-25 2008-05-29 Dolby Laboratories Licensing Corporation Multichannel Decorrelation In Spatial Audio Coding
US20080247556A1 (en) * 2007-02-21 2008-10-09 Wolfgang Hess Objective quantification of auditory source width of a loudspeakers-room system
US8238589B2 (en) * 2007-02-21 2012-08-07 Harman Becker Automotive Systems Gmbh Objective quantification of auditory source width of a loudspeakers-room system
AU2013263871B2 (en) * 2008-07-31 2015-07-30 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Signal generation for binaural signals
WO2010012478A3 (en) * 2008-07-31 2010-04-08 Fraunhofer-Gesellschaft zur Förderung der angewandten Forschung e.V. Signal generation for binaural signals
CN102172047A (en) * 2008-07-31 2011-08-31 弗劳恩霍夫应用研究促进协会 Signal generation for binaural signals
US9226089B2 (en) 2008-07-31 2015-12-29 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Signal generation for binaural signals
AU2009275418B9 (en) * 2008-07-31 2014-01-09 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Signal generation for binaural signals
AU2009275418B2 (en) * 2008-07-31 2013-12-19 Fraunhofer-Gesellschaft Zur Foerderung Der Angewandten Forschung E.V. Signal generation for binaural signals
US20100303246A1 (en) * 2009-06-01 2010-12-02 Dts, Inc. Virtual audio processing for loudspeaker or headphone playback
CN102597987A (en) * 2009-06-01 2012-07-18 Dts(英属维尔京群岛)有限公司 Virtual audio processing for loudspeaker or headphone playback
WO2010141371A1 (en) * 2009-06-01 2010-12-09 Dts, Inc. Virtual audio processing for loudspeaker or headphone playback
US8000485B2 (en) 2009-06-01 2011-08-16 Dts, Inc. Virtual audio processing for loudspeaker or headphone playback
US20120002024A1 (en) * 2010-06-08 2012-01-05 Lg Electronics Inc. Image display apparatus and method for operating the same
US8665321B2 (en) * 2010-06-08 2014-03-04 Lg Electronics Inc. Image display apparatus and method for operating the same
US9197979B2 (en) 2012-05-31 2015-11-24 Dts Llc Object-based audio system using vector base amplitude panning
US9264838B2 (en) 2012-12-27 2016-02-16 Dts, Inc. System and method for variable decorrelation of audio signals
US10679407B2 (en) 2014-06-27 2020-06-09 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for modeling interactive diffuse reflections and higher-order diffraction in virtual environment scenes
US20160034248A1 (en) * 2014-07-29 2016-02-04 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for conducting interactive sound propagation and rendering for a plurality of sound sources in a virtual environment scene
US9977644B2 (en) * 2014-07-29 2018-05-22 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for conducting interactive sound propagation and rendering for a plurality of sound sources in a virtual environment scene
KR20160026652A (en) * 2014-08-29 2016-03-09 후아웨이 테크놀러지 컴퍼니 리미티드 Sound signal processing method and apparatus
US9439019B2 (en) 2014-08-29 2016-09-06 Huawei Technologies Co., Ltd. Sound signal processing method and apparatus
KR101724514B1 (en) * 2014-08-29 2017-04-07 후아웨이 테크놀러지 컴퍼니 리미티드 Sound signal processing method and apparatus
EP2991382A1 (en) * 2014-08-29 2016-03-02 Huawei Technologies Co., Ltd. Sound signal processing method and apparatus
US11304020B2 (en) * 2016-05-06 2022-04-12 Dts, Inc. Immersive audio reproduction systems
US10248744B2 (en) 2017-02-16 2019-04-02 The University Of North Carolina At Chapel Hill Methods, systems, and computer readable media for acoustic classification and optimization for multi-modal rendering of real-world scenes
GB2565747A (en) * 2017-04-20 2019-02-27 Nokia Technologies Oy Enhancing loudspeaker playback using a spatial extent processed audio signal
GB2609667A (en) * 2021-08-13 2023-02-15 British Broadcasting Corp Audio rendering

Also Published As

Publication number Publication date
US8488796B2 (en) 2013-07-16

Similar Documents

Publication Publication Date Title
US8488796B2 (en) 3D audio renderer
US10820134B2 (en) Near-field binaural rendering
US9918179B2 (en) Methods and devices for reproducing surround audio signals
US10609503B2 (en) Ambisonic depth extraction
EP3550859B1 (en) Headphone virtualization
EP2997742B1 (en) An audio processing apparatus and method therefor
Gardner 3-D audio using loudspeakers
EP2920982A1 (en) Segment-wise adjustment of spatial audio signal to different playback loudspeaker setup
EP3569000B1 (en) Dynamic equalization for cross-talk cancellation
WO2009046460A2 (en) Phase-amplitude 3-d stereo encoder and decoder
CN111434126A (en) Signal processing device and method, and program
Jot et al. Binaural simulation of complex acoustic scenes for interactive audio
US20230143857A1 (en) Spatial Audio Reproduction by Positioning at Least Part of a Sound Field
US20220139403A1 (en) Audio System Height Channel Up-Mixing
TW202325047A (en) Apparatus, method or computer program for synthesizing a spatially extended sound source using variance or covariance data
TW202327379A (en) Apparatus, method or computer program for synthesizing a spatially extended sound source using modification data on a potentially modifying object

Legal Events

Date Code Title Description
AS Assignment

Owner name: CREATIVE TECHNOLOGY LTD, SINGAPORE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JOT, JEAN MARC;WALSH, MARTIN;PHILP, ADAM R;REEL/FRAME:020076/0408;SIGNING DATES FROM 20071019 TO 20071025

Owner name: CREATIVE TECHNOLOGY LTD, SINGAPORE

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:JOT, JEAN MARC;WALSH, MARTIN;PHILP, ADAM R;SIGNING DATES FROM 20071019 TO 20071025;REEL/FRAME:020076/0408

STCF Information on status: patent grant

Free format text: PATENTED CASE

FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: ENTITY STATUS SET TO SMALL (ORIGINAL EVENT CODE: SMAL); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YR, SMALL ENTITY (ORIGINAL EVENT CODE: M2552); ENTITY STATUS OF PATENT OWNER: SMALL ENTITY

Year of fee payment: 8