Sign In to Follow Application
View All Documents & Correspondence

Apparatus, Method And Computer Program For Encoding, Decoding, Scene Processing And Other Procedures Related To Dirac Based Spatial Audio Coding

Abstract: An apparatus for generating a description of a combined audio scene, comprises: an input interface (100) for receiving a first description of a first scene in a first format and a second description of a second scene in a second format, wherein the second format is different from the first format; a format converter (120) for converting the first description into a com-mon format and for converting the second description into the common format, when the second format is different from the common format; and a format combiner (140) for com-bining the first description in the common format and the second description in the common format to obtain the combined audio scene. Figure 1A is the representative figure.

Get Free WhatsApp Updates!
Notices, Deadlines & Correspondence

Patent Information

Application #
Filing Date
20 December 2021
Publication Number
09/2022
Publication Type
INA
Invention Field
ELECTRONICS
Status
Email
info@iprattorneys.com
Parent Application
Patent Number
Legal Status
Grant Date
2024-06-24
Renewal Date

Applicants

FRAUNHOFER-GESELLSCHAFT ZUR FÖRDERUNG DER ANGEWANDTEN FORSCHUNG E.V.
Hansastraße 27c, 80686 München, Germany.

Inventors

1. FUCHS, Guillaume
Joseph-Otto-Kolb-Str. 31, 91088 Bubenreuth, Germany.
2. HERRE, Jürgen
Rathsberger Str. 10a, 91054 Erlangen, Germany.
3. KÜCH, Fabian
Schützenweg 13, 91054 Erlangen, Germany.
4. DÖHLA, Stefan
Saidelsteig 61, 91058 Erlangen, Germany.
5. MULTRUS, Markus
Etzlaubweg 7, 90469 Nürnberg, Germany.
6. THIERGART, Oliver
Gebbertstraße 38, 91052 Erlangen, Germany.
7. WÜBBOLT, Oliver
Saarbrückener Str. 4, 30559 Hannover, Germany.
8. GHIDO, Florin
Am Stadtpark 21, 90409 Nürnberg, Germany.
9. BAYER, Stefan
Dortmunder Str. 14, 90425 Nürnberg, Germany.
10. JAEGERS, Wolfgang
Pfandlohe 4, 91301 Forchheim, Germany.

Specification

FORM 2
THE PATENTS ACT, 1970
(39 of 1970)
& The Patent Rules, 2003
COMPLETE SPECIFICATION
1.TITLE OF THE INVENTION:
APPARATUS, METHOD AND COMPUTER PROGRAM FOR ENCODING,
DECODING, SCENE PROCESSING AND OTHER PROCEDURES RELATED TO
DIRAC BASED SPATIAL AUDIO CODING
2. APPLICANT:
Name: FRAUNHOFER-GESELLSCHAFT ZUR FÖRDERUNG DER ANGEWANDTEN
FORSCHUNG E.V.
Nationality: Germany
Address: Hansastraße 27c, 80686 München, Germany.
3. PREAMBLE TO THE DESCRIPTION:
The following specification particularly describes the invention and the manner in which it
is to be performed:
2
Field of the Invention
The present invention is related to audio signal processing and particularly to audio signal
processing of audio descriptions 5 of audio scenes.
Introduction and state-of-the-art:
10 Transmitting an audio scene in three dimensions requires handling multiple channels which
usually engenders a large amount of data to transmit. Moreover 3D sound can be
represented in different ways: traditional channel-based sound where each transmission
channel is associated with a loudspeaker position; sound carried through audio objects,
which may be positioned in three dimensions independently of loudspeaker positions; and
15 scene-based (or Ambisonics), where the audio scene is represented by a set of coefficient
signals that are the linear weights of spatially orthogonal basis functions, e.g., spherical
harmonics. In contrast to channel-based representation, scene-based representation is
independent of a specific loudspeaker set-up, and can be reproduced on any loudspeaker
set-ups at the expense of an extra rendering process at the decoder.
20
For each of these formats, dedicated coding schemes were developed for efficiently storing
or transmitting at low bit-rates the audio signals. For example, MPEG surround is a
parametric coding scheme for channel-based surround sound, while MPEG Spatial Audio
Object Coding (SAOC) is a parametric coding method dedicated to object-based audio. A
25 parametric coding technique for high order of Ambisonics was also provided in the recent
standard MPEG-H phase 2.
In this context, where all three representations of the audio scene, channel-based, objectbased
and scene-based audio, are used and need to be supported, there is a need to design
30 a universal scheme allowing an efficient parametric coding of all three 3D audio
representations. Moreover there is a need to be able to encode, transmit and reproduce
complex audio scenes composed of a mixture of the different audio representations.
Directional Audio Coding (DirAC) technique [1] is an efficient approach to the analysis and
35 reproduction of spatial sound. DirAC uses a perceptually motivated representation of the
sound field based on direction of arrival (DOA) and diffuseness measured per frequency
band. It is built upon the assumption that at one time instant and at one critical band, the
spatial resolution of auditory system is limited to decoding one cue for direction and another
3
for inter-aural coherence. The spatial sound is then represented in frequency domain by
cross-fading two streams: a non-directional diffuse stream and a directional non-diffuse
stream.
DirAC was originally intended for recorded B-format sound but could 5 also serve as a
common format for mixing different audio formats. DirAC was already extended for
processing the conventional surround sound format 5.1 in [3]. It was also proposed to merge
multiple DirAC streams in [4]. Moreover, DirAC we extended to also support microphone
inputs other than B-format [6].
10
However, a universal concept is missing to make DirAC a universal representation of audio
scenes in 3D which also is able to support the notion of audio objects.
Few considerations were previously done for handling audio objects in DirAC. DirAC was
15 employed in [5] as an acoustic front end for the Spatial Audio Coder, SAOC, as a blind
source separation for extracting several talkers from a mixture of sources. It was, however,
not envisioned to use DirAC itself as the spatial audio coding scheme and to process directly
audio objects along with their metadata and to potentially combine them together and with
other audio representations.
20
It is an object of the present invention to provide an improved concept of handling and
processing audio scenes and audio scene descriptions.
This object is achieved by an apparatus for generating a description of a combined audio
25 scene of claim 1, a method of generating a description of a combined audio scene of claim
14, or a related computer program of claim 15.
Furthermore, this object is achieved by an apparatus for performing a synthesis of a plurality
of audio scenes of claim 16, a method for performing a synthesis of a plurality of audio
30 scenes of claim 20, or a related computer program in accordance with claim 21.
This object is furthermore achieved by an audio data converter of claim 22, a method for
performing an audio data conversion of claim 28, or a related computer program of claim
29.
35
Furthermore, this object is achieved by an audio scene encoder of claim 30, a method of
encoding an audio scene of claim 34, or a related computer program of claim 35.
4
Furthermore, this object is achieved by an apparatus for performing a synthesis of audio
data of claim 36, a method for performing a synthesis of audio data of claim 40, or a related
computer program of claim 41.
Embodiments of the invention relate to a universal parametric coding scheme 5 for 3D audio
scene built around the Directional Audio Coding paradigm (DirAC), a perceptually-motivated
technique for spatial audio processing. Originally DirAC was designed to analyze a B-format
recording of the audio scene. The present invention aims to extend its ability to process
efficiently any spatial audio formats such as channel-based audio, Ambisonics, audio
10 objects, or a mix of them
DirAC reproduction can easily be generated for arbitrary loudspeaker layouts and
headphones. The present invention also extends this ability to output additionally
Ambisonics, audio objects or a mix of a format. More importantly the invention enables the
15 possibility for the user to manipulate audio objects and to achieve, for example, dialogue
enhancement at the decoder end.
Context: System overview of a DirAC Spatial Audio Coder
20 In the following, an overview of a novel spatial audio coding system based on DirAC
designed for Immersive Voice and Audio Services (IVAS) is presented. The objective of
such a system is to be able to handle different spatial audio formats representing the audio
scene and to code them at low bit-rates and to reproduce the original audio scene as
faithfully as possible after transmission.
25
The system can accept as input different representations of audio scenes. The input audio
scene can be captured by multi-channel signals aimed to be reproduced at the different
loudspeaker positions, auditory objects along with metadata describing the positions of the
objects over time, or a first-order or higher-order Ambisonics format representing the sound
30 field at the listener or reference position.
Preferably the system is based on 3GPP Enhanced Voice Services (EVS) since the solution
is expected to operate with low latency to enable conversational services on mobile
networks.
35
Fig. 9 is the encoder side of the DirAC-based spatial audio coding supporting different audio
formats. As shown in Fig. 9, the encoder (IVAS encoder) is capable of supporting different
audio formats presented to the system separately or at the same time. Audio signals can
5
be acoustic in nature, picked up by microphones, or electrical in nature, which are supposed
to be transmitted to the loudspeakers. Supported audio formats can be multi-channel signal,
first-order and higher-order Ambisonics components, and audio objects. A complex audio
scene can also be described by combining different input formats. All audio formats are
then transmitted to the DirAC analysis 180, which extracts a parametric 5 representation of
the complete audio scene. A direction of arrival and a diffuseness measured per timefrequency
unit form the parameters. The DirAC analysis is followed by a spatial metadata
encoder 190, which quantizes and encodes DirAC parameters to obtain a low bit-rate
parametric representation.
10
Along with the parameters, a down-mix signal derived 160 from the different sources or
audio input signals is coded for transmission by a conventional audio core-coder 170. In
this case an EVS-based audio coder is adopted for coding the down-mix signal. The downmix
signal consists of different channels, called transport channels: the signal can be e.g.
15 the four coefficient signals composing a B-format signal, a stereo pair or a monophonic
down-mix depending of the targeted bit-rate. The coded spatial parameters and the coded
audio bitstream are multiplexed before being transmitted over the communication channel.
Fig. 10 is a decoder of the DirAC-based spatial audio coding delivering different audio
20 formats. In the decoder, shown in Fig. 10, the transport channels are decoded by the coredecoder
1020, while the DirAC metadata is first decoded 1060 before being conveyed with
the decoded transport channels to the DirAC synthesis 220, 240. At this stage (1040),
different options can be considered. It can be requested to play the audio scene directly on
any loudspeaker or headphone configurations as is usually possible in a conventional DirAC
25 system (MC in Fig. 10). In addition, it can also be requested to render the scene to
Ambisonics format for other further manipulations, such as rotation, reflection or movement
of the scene (FOA/HOA in Fig. 10). Finally, the decoder can deliver the individual objects
as they were presented at the encoder side (Objects in Fig. 10).
30 Audio objects could also be restituted but it is more interesting for the listener to adjust the
rendered mix by interactive manipulation of the objects. Typical object manipulations are
adjustment of level, equalization or spatial location of the object. Object-based dialogue
enhancement becomes, for example, a possibility given by this interactivity feature. Finally,
it is possible to output the original formats as they were presented at the encoder input. In
35 this case, it could be a mix of audio channels and objects or Ambisonics and objects. In
order to achieve separate transmission of multi-channels and Ambisonics components,
several instances of the described system could be used.
6
The present invention is advantageous in that, particularly in accordance with the first
aspect, a framework is established in order to combine different scene descriptions into a
combined audio scene by way of a common format, that allows to combine the different
audio scene descriptions.
5
This common format may, for example, be the B-format or may be the pressure/velocity
signal representation format, or can, preferably, also be the DirAC parameter representation
format.
10 This format is a compact format that, additionally, allows a significant amount of user
interaction on the one hand and that is, on the other hand, useful with respect to a required
bitrate for representing an audio signal.
In accordance with a further aspect of the present invention, a synthesis of a plurality of
15 audio scenes can be advantageously performed by combing two or more different DirAC
descriptions. Both these different DirAC descriptions can be processed by combining the
scenes in the parameter domain or, alternatively, by separately rendering each audio scene
and by then combining the audio scenes that have been rendered from the individual DirAC
descriptions in the spectral domain or, alternatively, already in the time domain.
20
This procedure allows for a very efficient and nevertheless high quality processing of
different audio scenes that are to be combined into a single scene representation and,
particularly, a single time domain audio signal.
25 A further aspect of the invention is advantageous in that a particularly useful audio data
converted for converting object metadata into DirAC metadata is derived where this audio
data converter can be used in the framework of the first, the second or the third aspect or
can also be applied independent from each other. The audio data converter allows
efficiently converting audio object data, for example, a waveform signal for an audio object,
30 and corresponding position data, typically, with respect to time for representing a certain
trajectory of an audio object within a reproduction setting up into a very useful and compact
audio scene description, and, particularly, the DirAC audio scene description format. While
a typical audio object description with an audio object waveform signal and an audio object
position metadata is related to a particular reproduction setup or, generally, is related to a
35 certain reproduction coordinate system, the DirAC description is particularly useful in that it
is related to a listener or microphone position and is completely free of any limitations with
respect to a loudspeaker setup or a reproduction setup.
7
Thus, the DirAC description generated from audio object metadata signals additionally
allows for a very useful and compact and high quality combination of audio objects different
from other audio object combination technologies such as spatial audio object coding or
amplitude panning of objects in a reproduction setup.
5
An audio scene encoder in accordance with a further aspect of the present invention is
particularly useful in providing a combined representation of an audio scene having DirAC
metadata and, additionally, an audio object with audio object metadata.
10 Particularly, in this situation, it is particularly useful and advantageous for a high interactivity
in order to generate a combined metadata description that has DirAC metadata on the one
hand and, in parallel, object metadata on the other hand. Thus, in this aspect, the object
metadata is not combined with the DirAC metadata, but is converted into DirAC-like
metadata so that the object metadata comprises at direction or, additionally, a distance
15 and/or a diffuseness of the individual object together with the object signal. Thus, the object
signal is converted into a DirAC-like representation so that a very flexible handling of a
DirAC representation for a first audio scene and an additional object within this first audio
scene is allowed and made possible. Thus, for example, specific objects can be very
selectively processed due to the fact that their corresponding transport channel on the one
20 hand and DirAC-style parameters on the other hand are still available.
In accordance with a further aspect of the invention, an apparatus or method for performing
a synthesis of audio data are particularly useful in that a manipulator is provided for
manipulating a DirAC description of one or more audio objects, a DirAC description of the
25 multichannel signal or a DirAC description of first order Ambisonics signals or higher
Ambisonics signals. And, the manipulated DirAC description is then synthesized using a
DirAC synthesizer.
This aspect has the particular advantage that any specific manipulations with respect to any
30 audio signals are very usefully and efficiently performed in the DirAC domain, i.e., by
manipulating either the transport channel of the DirAC description or by alternatively
manipulating the parametric data of the DirAC description. This modification is substantially
more efficient and more practical to perform in the DirAC domain compared to the
manipulation in other domains. Particularly, position-dependent weighting operations as
35 preferred manipulation operations can be particularly performed in the DirAC domain. Thus,
in a specific embodiment, a conversion of a corresponding signal representation in the
DirAC domain and, then, performing the manipulation within the DirAC domain is a
8
particularly useful application scenario for modern audio scene processing and
manipulation.
Preferred embodiments are subsequently discussed with respect to their accompanying
drawings, 5 in which:
Fig. 1a is a block diagram of a preferred implementation of an apparatus or method
for generating a description of a combined audio scene in accordance with a
first aspect of the invention;
10
Fig. 1b is an implementation of the generation of a combined audio scene, where
the common format is the pressure/velocity representation;
Fig. 1c is a preferred implementation of the generation of a combined audio scene,
15 where the DirAC parameters and the DirAC description is the common
format;
Fig. 1d is a preferred implementation of the combiner in Fig. 1c illustrating two
different alternatives for the implementation of the combiner of DirAC
20 parameters of different audio scenes or audio scene descriptions;
Fig. 1e is a preferred implementation of the generation of a combined audio scene
where the common format is the B-format as an example for an Ambisonics
representation;
25
Fig. 1f is an illustration of an audio object/DirAC converter useful in the context of,
of example, Fig. 1c or 1d or useful in the context of the third aspect relating
to a metadata converter;
30 Fig. 1g is an exemplary illustration of a 5.1 multichannel signal into a DirAC
description;
Fig. 1h is a further illustration the conversion of a multichannel format into the DirAC
format in the context of an encoder and a decoder side;
35
Fig. 2a illustrates an embodiment of an apparatus or method for performing a
synthesis of a plurality of audio scenes in accordance with a second aspect
of the present invention;
9
Fig. 2b illustrates a preferred implementation of the DirAC synthesizer of Fig. 2a;
Fig. 2c illustrates a further implementation of the DirAC synthesizer with a
combination of 5 rendered signals;
Fig. 2d illustrates an implementation of a selective manipulator either connected
before the scene combiner 221 of Fig. 2b or before the combiner 225 of Fig.
2c;
10
Fig. 3a is a preferred implementation of an apparatus or method for performing and
audio data conversion in accordance with a third aspect of the present
invention;
15 Fig. 3b is a preferred implementation of the metadata converter also illustrated in
Fig. 1f;
Fig. 3c is a flowchart for performing a further implementation of a audio data
conversion via the pressure/velocity domain;
20
Fig. 3d illustrates a flowchart for performing a combination within the DirAC domain;
Fig. 3e illustrates a preferred implementation for combining different DirAC
descriptions, for example as illustrated in Fig. 1d with respect to the first
25 aspect of the present invention;
Fig. 3f illustrates the conversion of an object position data into a DirAC parametric
representation;
30 Fig. 4a illustrates a preferred implementation of an audio scene encoder in
accordance with a fourth aspect of the present invention for generating a
combined metadata description comprising the DirAC metadata and the
object metadata;
35 Fig. 4b illustrates a preferred embodiment with respect to the fourth aspect of the
present invention;
10
Fig. 5a illustrates a preferred implementation of an apparatus for performing a
synthesis of audio data or a corresponding method in accordance with a fifth
aspect of the present invention;
Fig. 5b illustrates a preferred implementation of the DirAC synthesizer 5 of Fig. 5a;
Fig. 5c illustrates a further alternative of the procedure of the manipulator of Fig. 5a;
Fig. 5d illustrates a further procedure for the implementation of the Fig. 5a
10 manipulator;
Fig. 6 illustrates an audio signal converter for generating, from a mono-signal and
a direction of arrival information, i.e., from an exemplary DirAC description,
where the diffuseness is, for example, set to zero, a B-format representation
15 comprising an omnidirectional component and directional components in X,
Y and Z directions;
Fig. 7a illustrates an implementation of a DirAC analysis of a B-Format microphone
signal;
20
Fig. 7b illustrates an implementation of a DirAC synthesis in accordance with a
known procedure;
Fig. 8 illustrates a flowchart for illustrating further embodiments of, particularly, the
25 Fig. 1a embodiment;
Fig. 9 is the encoder side of the DirAC-based spatial audio coding supporting
different audio formats;
30 Fig. 10 is a decoder of the DirAC-based spatial audio coding delivering different
audio formats;
Fig. 11 is a system overview of the DirAC-based encoder/decoder combining
different input formats in a combined B-format;
35
Fig. 12 is a system overview of the DirAC-based encoder/decoder combining in the
pressure/velocity domain;
11
Fig. 13 is a system overview of the DirAC-based encoder/decoder combining
different input formats in the DirAC domain with the possibility of object
manipulation at the decoder side;
Fig. 14 is a system overview of the DirAC-based encoder/5 decoder combining
different input formats at the decoder-side through a DirAC metadata
combiner;
Fig. 15 is a system overview of the DirAC-based encoder/decoder combining
10 different input formats at the decoder-side in the DirAC synthesis; and
Fig. 16a-f illustrates several representations of useful audio formats in the context of
the first to fifth aspects of the present invention.
15 Fig. 1a illustrates a preferred embodiment of an apparatus for generating a description of a
combined audio scene. The apparatus comprises an input interface 100 for receiving a first
description of a first scene in a first format and a second description of a second scene in a
second format, wherein the second format is different from the first format. The format can
be any audio scene format such as any of the formats or scene descriptions illustrated from
20 Figs. 16a to 16f.
Fig. 16a, for example, illustrates an object description consisting, typically, of a (encoded)
object 1 waveform signal such as a mono-channel and corresponding metadata related to
the position of object 1, where this is information is typically given for each time frame or a
25 group of time frames, and which the object 1 waveforms signal is encoded. Corresponding
representations for a second or further object can be included as illustrated in Fig. 16a.
Another alternative can be an object description consisting of an object downmix being a
mono-signal, a stereo-signal with two channels or a signal with three or more channels and
30 related object metadata such as object energies, correlation information per time/frequency
bin and, optionally, the object positions. However, the object positions can also be given at
the decoder side as typical rendering information and, therefore, can be modified by a user.
The format in Fig. 16b can, for example, be implemented as the well-known SAOC (spatial
audio object coding) format.
35
Another description of a scene is illustrated in Fig. 16c as a multichannel description having
an encoded or non-encoded representation of a first channel, a second channel, a third
channel, a fourth channel, or a fifth channel, where the first channel can be the left channel
12
L, the second channel can be the right channel R, the third channel can be the center
channel C, the fourth channel can be the left surround channel LS and the fifth channel can
be the right surround channel RS. Naturally, the multichannel signal can have a smaller or
higher number of channels such as only two channels for a stereo channel or six channels
for a 5.1 format or eight channels for a 5 7.1 format, etc.
A more efficient representation of a multichannel signal is illustrated in Fig. 16d, where the
channel downmix such as a mono downmix, or stereo downmix or a downmix with more
than two channels is associated with parametric side information as channel metadata for,
10 typically, each time and/or frequency bin. Such a parametric representation can, for
example, be implemented in accordance with the MPEG surround standard.
Another representation of an audio scene can, for example, be the B-format consisting of
an omnidirectional signal W, and directional components X, Y, Z as shown in Fig. 16e. This
15 would be a first order or FoA signal. A higher order Ambisonics signal, i.e., an HoA signal
can have additional components as is known in the art.
The Fig. 16e representation is, in contrast to the Fig. 16c and Fig. 16d representation a
representation that is non-dependent on a certain loudspeaker set up, but describes a
20 sound field as experienced at a certain (microphone or listener) position.
Another such sound field description is the DirAC format as, for example, illustrated in Fig.
16f. The DirAC format typically comprises a DirAC downmix signal which is a mono or stereo
or whatever downmix signal or transport signal and corresponding parametric side
25 information. This parametric side information is, for example, a direction of arrival
information per time/frequency bin and, optionally, diffuseness information per
time/frequency bin.
The input into the input interface 100 of Fig. 1a can be, for example, in any one of those
30 formats illustrated with respect to Fig. 16a to Fig. 16f. The input interface 100 forwards the
corresponding format descriptions to a format converter 120. The format converter 120 is
configured for converting the first description into a common format and for converting the
second description into the same common format, when the second format is different from
the common format. When, however, the second format is already in the common format,
35 then the format converter only convers the first description into the common format, since
the first description is in a format different from the common format.
13
Thus, at the output of the format converter or, generally, at the input of a format combiner,
there does exist a representation of the first scene in the common format and the
representation of the second scene in the same common format. Due to the fact that both
descriptions are now included in one and the same common format, the format combiner
can now combine the first description and the second description to 5 obtain a combined
audio scene.
In accordance with an embodiment illustrated in Fig. 1e, the format converter 120 is
configured to convert the first description into a first B-format signal as, for example,
10 illustrated at 127 in Fig. 1e and to compute the B-format representation for the second
description as illustrated in Fig. 1e at 128.
Then, the format combiner 140 is implemented as a component signal adder illustrated at
146a for the W component adder, 146b for the X component adder, illustrated at 146c for
15 the Y component adder and illustrated at 146d for the Z component adder.
Thus, in the Fig. 1e embodiment, the combined audio scene can be a B-format
representation and the B-format signals can then operate as the transport channels and
can then be encoded via a transport channel encoder 170 of Fig. 1a. Thus, the combined
20 audio scene with respect to B-format signal can be directly input into the encoder 170 of
Fig. 1a to generate an encoded B-format signal that could then be output via the output
interface 200. In this case, any spatial metadata are not required, but, at the price of an
encoded representation of four audio signals, i.e., the omnidirectional component W and
the directional components X, Y, Z.
25
Alternatively, the common format is the pressure/velocity format as illustrated in Fig. 1b. To
this end, the format converter 120 comprises a time/frequency analyzer 121 for the first
audio scene and the time/frequency analyzer 122 for the second audio scene or, generally,
the audio scene with number N, where N is an integer number.
30
Then, for each such spectral representation generated by the spectral converters 121, 122,
pressure and velocity are computed as illustrated at 123 and 124, and, the format combiner
then is configured to calculate a summed pressure signal on the one hand by summing the
corresponding pressure signals generated by the blocks 123, 124. And, additionally, an
35 individual velocity signal is calculated as well by each of the blocks 123, 124 and the velocity
signals can be added together in order to obtain a combined pressure/velocity signal.
14
Depending on the implementation, the procedures in blocks 142, 143 does not necessarily
have to be performed. Instead, the combined or “summed” pressure signal and the
combined or “summed” velocity signal can be encoded in an analogy as illustrated in Fig.
1e of the B-format signal and this pressure/velocity representation could be encoded while
once again via that encoder 170 of Fig. 1a and could then be transmitted 5 to the decoder
without any additional side information with respect to spatial parameters, since the
combined pressure/velocity representation already includes the necessary spatial
information for obtaining a finally rendered high quality sound field on a decoder-side .
10 In an embodiment, however, it is preferred to perform a DirAC analysis to the
pressure/velocity representation generated by block 141. To this end, the intensity vector
142 is calculated and, in block 143, the DirAC parameters from the intensity vector is
calculated, and, then, the combined DirAC parameters are obtained as a parametric
representation of the combined audio scene. To this end, the DirAC analyzer 180 of Fig. 1a
15 is implemented to perform the functionality of block 142 and 143 of Fig. 1b. And, preferably,
the DirAC data is additionally subjected to a metadata encoding operation in metadata
encoder 190. The metadata encoder 190 typically comprises a quantizer and entropy coder
in order to reduce the bitrate required for the transmission of the DirAC parameters.
20 Together with the encoded DirAC parameters, an encoded transport channel is also
transmitted. The encoded transport channel is generated by the transport channel generator
160 of Fig. 1a that can, for example, be implemented as illustrated in Fig. 1b by a first
downmix generator 161 for generating a downmix from the first audio scene and a N-th
downmix generator 162 for generating a downmix from the N-th audio scene.
25
Then, the downmix channels are combined in combiner 163 typically by a straightforward
addition and the combined downmix signal is then the transport channel that is encoded by
the encoder 170 of Fig. 1a. The combined downmix can, for example, be a stereo pair, i.e.,
a first channel and a second channel of a stereo representation or can be a mono channel,
30 i.e., a single channel signal.
In accordance with a further embodiment illustrated in Fig. 1c, a format conversion in the
format converter 120 is done to directly convert each of the input audio formats into the
DirAC format as the common format. To this end, the format converter 120 once again forms
35 a time-frequency conversion or a time/frequency analysis in corresponding blocks 121 for
the first scene and block 122 for a second or further scene. Then, DirAC parameters are
derived from the spectral representations of the corresponding audio scenes illustrated at
125 and 126. The result of the procedure in blocks 125 and 126 are DirAC parameters
15
consisting of energy information per time/frequency tile, a direction of arrival information
eDOA per time/frequency tile and a diffuseness information ψ for each time/frequency tile.
Then, the format combiner 140 is configured to perform a combination directly in the DirAC
parameter domain in order to generate combined DirAC parameters ψ for the diffuseness
and eDOA for the direction of arrival. Particularly, the energy information 5 E1 and EN are
required by the combiner 144 but are not part of the final combined parametric
representation generated by the format combiner 140.
Thus, comparing Fig. 1c to Fig. 1e reveals that, when the format combiner 140 already
10 performs a combination in the DirAC parameter domain, the DirAC analyzer 180 is not
necessary and not implemented. Instead, the output of the format combiner 140 being the
output of block 144 in Fig. 1c is directly forwarded to the metadata encoder 190 of Fig. 1a
and from there into the output interface 200 so that the encoded spatial metadata and,
particularly, the encoded combined DirAC parameters are included in the encoded output
15 signal output by the output interface 200.
Furthermore, the transport channel generator 160 of Fig. 1a may receive, already from the
input interface 100, a waveform signal representation for the first scene and the waveform
signal representation for the second scene. These representations are input into the
20 downmix generator blocks 161, 162 and the results are added in block 163 to obtain a
combined downmix as illustrated with respect to Fig. 1b.
Fig. 1d illustrates a similar representation with respect to Fig. 1c. However, in Fig. 1d, the
audio object waveform is input into the time/frequency representation converter 121 for
25 audio object 1 and 122 for audio object N. Additionally, the metadata are input, together
with the spectral representation into the DirAC parameter calculators 125, 126 as illustrated
also in Fig. 1c.
However, Fig. 1d provides a more detailed representation with respect to how preferred
30 implementations of the combiner 144 operate. In a first alternative, the combiner performs
an energy-weighted addition of the individual diffuseness for each individual object or scene
and, a corresponding energy-weighted calculation of a combined DoA for each
time/frequency tile is performed as illustrated in the lower equation of alternative 1.
35 However, other implementations can be performed as well. Particularly, another very
efficient calculation is set the diffuseness to zero for the combined DirAC metadata and to
select, as the direction of arrival for each time/frequency tile the direction of arrival
calculated from a certain audio object that has the highest energy within the specific
16
time/frequency tile. Preferably, the procedure in Fig. 1d is more appropriate when the input
into the input interface are individual audio objects correspondingly represented a waveform
or mono-signal for each object and corresponding metadata such as position information
illustrated with respect to Fig. 16a or 16b.
5
However, in the Fig. 1c embodiment, the audio scene may be any other of the
representations illustrated in Fig. 16c, 16d, 16e or 16f. Then, there can be metadata or not,
i.e., the metadata in Fig. 1c is optional. Then, however, a typically useful diffuseness is
calculated for a certain scene description such as an Ambisonics scene description in Fig.
10 16e and, then, the first alternative of the way how the parameters are combined is preferred
over the second alternative of Fig. 1d. Therefore, in accordance with the invention, the
format converter 120 is configured to convert a high order Ambisonics or a first order
Ambisonics format into the B-format, wherein the high order Ambisonics format is truncated
before being converted into the B-format.
15
In a further embodiment, the format converter is configured to project an object or a channel
on spherical harmonics at the reference position to obtain projected signals, and wherein
the format combiner is configured to combine the projection signals to obtain B-format
coefficients, wherein the object or the channel is located in space at a specified position
20 and has an optional individual distance from a reference position. This procedure
particularly works well for the conversion of object signals or multichannel signals into first
order or high order Ambisonics signals.
In a further alternative, the format converter 120 is configured to perform a DirAC analysis
25 comprising a time-frequency analysis of B-format components and a determination of
pressure and velocity vectors and where the format combiner is then configured to combine
different pressure/velocity vectors and where the format combiner further comprises the
DirAC analyzer 180 for deriving DirAC metadata from the combined pressure/velocity data.
30 In a further alternative embodiment, the format converter is configured to extract the DirAC
parameters directly from the object metadata of an audio object format as the first or second
format, where the pressure vector for the DirAC representation is the object waveform signal
and the direction is derived from the object position in space or the diffuseness is directly
given in the object metadata or is set to a default value such as the zero value.
35
In a further embodiment, the format converter is configured to convert the DirAC parameters
derived from the object data format into pressure/velocity data and the format combiner is
17
configured to combine the pressure/velocity data with pressure/velocity data derived from
different description of one or more different audio objects.
However, in a preferred implementation illustrated with respect to Fig. 1c and 1d, the format
combiner is configured to directly combine the DirAC parameters derived 5 by the format
converter 120 so that the combined audio scene generated by block 140 of Fig. 1a is already
the final result and a DirAC analyzer 180 illustrated in Fig. 1a is not necessary, since the
data output by the format combiner 140 is already in the DirAC format.
10 In a further implementation, the format converter 120 already comprises a DirAC analyzer
for first order Ambisonics or a high order Ambisonics input format or a multichannel signal
format. Furthermore, the format converter comprises a metadata converter for converting
the object metadata into DirAC metadata, and such a metadata converter is, for example,
illustrated in Fig. 1f at 150 that once again operates on the time/frequency analysis in block
15 121 and calculates the energy per band per time frame illustrated at 147, the direction of
arrival illustrated at block 148 of Fig. 1f and the diffuseness illustrated at block 149 of Fig.
1f. And, the metadata are combined by the combiner 144 for combining the individual DirAC
metadata streams, preferably by a weighted addition as illustrated exemplarily by one of the
two alternatives of the Fig. 1d embodiment.
20
Multichannel channel signals can be directly converted to B-format. The obtained B-format
can be then processed by a conventional DirAC. Fig. 1g illustrates a conversion 127 to Bformat
and a subsequent DirAC processing 180.
25 Reference [3] outlines ways to perform the conversion from multi-channel signal to Bformat.
In principle, converting multi-channel audio signals to B-format is simple: virtual
loudspeakers are defined to be at different positions of the loudspeaker layout. For example
for 5.0 layout, loudspeakers are positioned on the horizontal plane at azimuth angles +/-30
and +/-110 degrees. A virtual B-format microphone is then defined to be in the center of the
30 loudspeakers, and a virtual recording is performed. Hence, the W channel is created by
summing all loudspeaker channels of the 5.0 audio file. The process for getting W and other
B-format coefficients can be then summarized:
𝑊 = 􀷍 􀶨
1
2
𝑤􀯜𝑠􀯜
􀯞
􀯜􀭀􀬵
𝑋 = 􀷍 𝑤􀯜𝑠􀯜(cos(𝜃􀯜) cos(𝜑􀯜 ))
􀯞
􀯜􀭀􀬵
35
18
𝑌 = 􀷍 𝑤􀯜𝑠􀯜 (sin(𝜃􀯜) cos(𝜑􀯜 ))
􀯞
􀯜􀭀􀬵
𝑍 = 􀷍 𝑤􀯜𝑠􀯜 (sin(𝜑􀯜))
􀯞
􀯜􀭀􀬵
where 𝑠􀯜 are the multichannel signals located in the space at the loudspeaker positions
defined by the azimuth angle 𝜃􀯜 and elevation angle 𝜑􀯜 , of each loudspeaker 5 and 𝑤􀯜 are
weights function of the distance. If the distance is not available or simply ignored, then 𝑤􀯜 =
1. Though, this simple technique is limited since it is an irreversible process. Moreover since
the loudspeaker are usually distributed non-uniformly, there is also a bias in the estimation
done by a subsequent DirAC analysis towards the direction with the highest loudspeaker
10 density. For example in 5.1 layout, there will be a bias towards the front since there are
more loudspeakers in the front than in the back.
To address this issue, a further technique was proposed in [3] for processing 5.1
multichannel signal with DirAC. The final coding scheme will then look as illustrated in Fig.
15 1h showing the B-format converter 127, the DirAC analyzer 180 as generally described with
respect to element 180 in Fig. 1, and the other elements 190, 1000, 160, 170, 1020, and/or
220, 240.
In a further embodiment, the output interface 200 is configured to add, to the combined
20 format, a separate object description for an audio object, where the object description
comprises at least one of a direction, a distance, a diffuseness or any other object attribute,
where this object has a single direction throughout all frequency bands and is either static
or moving slower than a velocity threshold.
25 This feature is furthermore elaborated in more detail with respect to the fourth aspect of the
present invention discussed with respect to Fig. 4a and Fig. 4b.
1st Encoding Alternative: Combining and processing different audio representations
through B-format or equivalent representation.
30
A first realization of the envisioned encoder can be achieved by converting all input format
into a combined B-format as it is depicted in Fig. 11.
Fig. 11: System overview of the DirAC-based encoder/decoder combining different input
35 formats in a combined B-format
19
Since DirAC is originally designed for analyzing a B-format signal, the system converts the
different audio formats to a combined B-format signal. The formats are first individually
converted 120 into a B-format signal before being combined together by summing their Bformat
components W,X,Y,Z. First Order Ambisonics (FOA) components 5 can be normalized
and re-ordered to a B-format. Assuming FOA is in ACN/N3D format, the four signals of the
B-format input are obtained by:

⎪ ⎪ ⎪ ⎪

⎪ ⎪ ⎪ ⎪

𝑊 = 𝑌􀬴
􀬴
𝑋 = 􀶨
2
3
𝑌􀬵
􀬵
𝑌 = 􀶨
2
3
𝑌􀬵
􀬿􀬵
𝑍 = 􀶨
2
3
𝑌􀬵
􀬴
10
Where 𝑌􀯠 􀯟
denotes the Ambisonics component of order 𝑙 and index 𝑚, −𝑙 ≤ 𝑚 ≤ +𝑙. Since
FOA components are fully contained in higher order Ambisonics format, HOA format needs
only to be truncated before being converted into B-format.
15 Since objects and channels have determined positions in the space, it is possible to project
each individual object and channel on spherical Harmonics (SH) at the center position such
as recording or reference position. The sum of the projections allows combining different
objects and multiple channels in a single B-format and can be then processed by the DirAC
analysis. The B-format coefficients (W,X,Y,Z) are then given by:
𝑊 = 􀷍 􀶨
1
2
𝑤􀯜𝑠􀯜
􀯞
􀯜􀭀􀬵
20
𝑋 = 􀷍 𝑤􀯜𝑠􀯜(cos(𝜃􀯜) cos(𝜑􀯜 ))
􀯞
􀯜􀭀􀬵
𝑌 = 􀷍 𝑤􀯜𝑠􀯜 (sin(𝜃􀯜) cos(𝜑􀯜 ))
􀯞
􀯜􀭀􀬵
𝑍 = 􀷍 𝑤􀯜𝑠􀯜 (sin(𝜑􀯜))
􀯞
􀯜􀭀􀬵
25 where 𝑠􀯜 are independent signals located in the space at positions defined by the azimuth
angle 𝜃􀯜 and elevation angle 𝜑􀯜 , and 𝑤􀯜 are weights function of the distance. If the distance
is not available or simply ignored, then 𝑤􀯜 = 1. For example, the independent signals can
20
correspond to audio objects that are located at the given position or the signal associated
with a loudspeaker channel at the specified position.
In applications where an Ambisonics representation of orders higher than first order is
desired, the Ambisonics coefficients generation presented above for first 5 order is extended
by additionally considering higher-order components.
The transport channel generator 160 can directly receive the multichannel signal, objects
waveform signals, and the higher order Ambisonics components. The transport channel
10 generator will reduce the number of input channels to transmit by downmixing them. The
channels can be mixed together as in MPEG surround in a mono or stereo downmix, while
object waveform signals can be summed up in a passive way into a mono downmix. In
addition, from the higher order Ambisonics, it is possible to extract a lower order
representation or to create by beamforming a stereo downmix or any other sectioning of the
15 space. If the downmixes obtained from the different input format are compatible with each
other, they can be combined together by a simple addition operation.
Alternatively, the transport channel generator 160 can receive the same combined B-format
as that conveyed to the DirAC analysis. In this case, a subset of the components or the
20 result of a beamforming (or other processing) form the transport channels to be coded and
transmitted to the decoder. In the proposed system, a conventional audio coding is required
which can be based on, but is not limited to, the standard 3GPP EVS codec. 3GPP EVS is
the preferred codec choice because of its ability to code either speech or music signals at
low bit-rates with high quality while requiring a relatively low delay enabling real-time
25 communications.
At a very low bit-rate, the number of channels to transmit needs to be limited to one and
therefore only the omnidirectional microphone signal W of the B-format is transmitted. If bitrate
allows, the number of transport channels can be increased by selecting a subset of the
30 B-format components. Alternatively, the B-format signals can be combined into a
beamformer 160 steered to specific partitions of the space. As an example two cardioids
can be designed to point at opposite directions, for example to the left and the right of the
spatial scene:
􀵜
𝐿 = √2𝑊 + 𝑌
𝑅 = √2𝑊 − 𝑌
35
These two stereo channels L and R can be then efficiently coded 170 by a joint stereo
coding. The two signals will be then adequately exploited by the DirAC Synthesis at the
21
decoder side for rendering the sound scene. Other beamforming can be envisioned, for
example a virtual cardioid microphone can be pointed toward any directions of given
azimuth 𝜃 and elevation 𝜑:
𝐶 = √2W + cos(𝜃) cos(𝜑) 𝑋 + sin(5 𝜃) cos(𝜑) 𝑌 + sin (𝜑)𝑍
Further ways of forming transmission channels can be envisioned that carry more spatial
information than a single monophonic transmission channel would do.
Alternatively, the 4 coefficients of the B-format can be directly transmitted. In that case the
10 DirAC metadata can be extracted directly at the decoder side, without the need of
transmitting extra information for the spatial metadata.
Fig.12 shows another alternative method for combining the different input formats. Fig. 12
also is a system overview of the DirAC-based encoder/decoder combining in
15 Pressure/velocity domain.
Both multichannel signal and Ambisonics components are input to a DirAC analysis 123,
124. For each input format a DirAC analysis is performed consisting of a time-frequency
analysis of the B-format components 𝑤􀯜(𝑛), 𝑥􀯜 (𝑛), 𝑦􀯜(𝑛), 𝑧􀯜 (𝑛) and the determination of the
20 pressure and velocity vectors:
𝑃􀯜(𝑛, 𝑘) = 𝑊􀯜 (𝑘, 𝑛)
𝑈􀯜(𝑛, 𝑘) = 𝑋􀯜 (𝑘, 𝑛)𝒆𝒙 + 𝑌􀯜(𝑘, 𝑛)𝒆𝒚 + 𝑍􀯜 (𝑘, 𝑛)𝒆𝒛
25 where 𝑖 is the index of the input and, 𝑘 and 𝑛 time and frequency indices of the timefrequency
tile, and 𝒆𝒙, 𝒆𝒚, 𝒆𝒛 represent the Cartesian unit vectors.
𝑃(𝑛, 𝑘) and 𝑈(𝑛, 𝑘) are necessary to compute the DirAC parameters, namely DOA and
diffuseness. The DirAC metadata combiner can exploit that 𝑁 sources which play together
30 result in a linear combination of their pressures and particle velocities that would be
measured when they are played alone. The combined quantities are then derived by:
𝑃(𝑛, 𝑘) = 􀷍 𝑃􀯜 (𝑛, 𝑘)
􀯇
􀯜􀭀􀬵
𝑈(𝑛, 𝑘) = 􀷍 𝑈􀯜(𝑛, 𝑘)
􀯇
􀯜􀭀􀬵
22
The combined DirAC parameters are computed 143 through the computation of the
combined intensity vector:
𝐼(𝑘, 𝑛) =
􀬵
􀬶
ℜ􀵛𝑃(𝑘, 𝑛). 􀴤𝑈􀴤􀴤(􀴤𝑘􀴤􀴤,􀴤𝑛􀴤􀴤)􀵟,
5
where (􀴤􀴤.􀴤􀴤) denotes complex conjugation. The diffuseness of the combined sound field is
given by:
𝜓(𝑘, 𝑛) = 1 −
‖Ε{𝐼(𝑘, 𝑛)}‖
𝑐Ε{𝐸(𝑘, 𝑛)}
10 where Ε{. } denotes the temporal averaging operator, 𝑐 the speed of sound and 𝐸(𝑘, 𝑛) the
sound field energy given by:
𝐸(𝑘, 𝑛) =
𝜌􀬴
4
‖𝑈(𝑘, 𝑛)‖􀬶 +
1
𝜌􀬴𝑐􀬶 |𝑃(𝑘, 𝑛)|􀬶
15 The direction of arrival (DOA) is expressed by means of the unit vector 𝑒􀮽􀯈􀮺(𝑘, 𝑛), defined
as
𝑒􀮽􀯈􀮺(𝑘, 𝑛) = −
𝐼(𝑘, 𝑛)
‖𝐼(𝑘, 𝑛)‖
20 If an audio object is input, the DirAC parameters can be directly extracted from the object
metadata while the pressure vector 𝑃􀯜 (𝑘, 𝑛) is the object essence (waveform) signal. More
precisely, the direction is straightforwardly derived from the object position in the space,
while the diffuseness is directly given in the object metadata or - if not available – can be
set by default to zero. From the DirAC parameters the pressure and the velocity vectors are
25 directly given by:
𝑃 􀷠
􀯜 (𝑘, 𝑛) = 􀶧1 − 𝜓􀯜 (𝑘, 𝑛)𝑃􀯜(𝑘, 𝑛)
𝑈 􀷡
􀯜(𝑘, 𝑛) = −
1
𝜌􀬴𝑐
𝑃 􀷠
􀯜 (𝑘, 𝑛). 𝑒􀮽􀯈􀮺
􀯜 (𝑘, 𝑛)
30
The combination of objects or the combination of an object with different input formats is
then obtained by summing the pressure and velocity vectors as explained previously.
23
In summary, the combination of different input contributions (Ambisonics, channels, objects)
is performed in the pressure / velocity domain and the result is then subsequently converted
into direction / diffuseness DirAC parameters. Operating in pressure/velocity domain is the
theoretically equivalent to operate in B-format. The main benefit of this alternative compared
to the previous one is the possibility to optimize the DirAC analysis according 5 to each input
format as it is proposed in [3] for surround format 5.1.
The main drawback of such a fusion in a combined B-format or pressure/velocity domain is
that the conversion happening at the front-end of the processing chain is already a
10 bottleneck for the whole coding system. Indeed, converting audio representations from
higher-order Ambisonics, objects or channels to a (first-order) B-format signal engenders
already a great loss of spatial resolution which cannot be recovered afterwards.
2st Encoding Alternative: combination and processing in DirAC domain
15
To circumvent the limitations of converting all input formats into a combined B-format signal,
the present alternative proposes to derive the DirAC parameters directly from the original
format and then to combine them subsequently in the DirAC parameter domain. The general
overview of such a system is given in Fig. 13. Fig. 13 is a system overview of the DirAC20
based encoder/decoder combining different input formats in DirAC domain with the
possibility of object manipulation at the decoder side.
In the following, we can also consider individual channels of a multichannel signal as an
audio object input for the coding system. The object metadata is then static over time and
25 represent the loudspeaker position and distance related to listener position.
The objective of this alternative solution is to avoid the systematic combination of the
different input formats into to a combined B-format or equivalent representation. The aim is
to compute the DirAC parameters before combining them. The method avoids then any
30 biases in the direction and diffuseness estimation due to the combination. Moreover, it can
optimally exploit the characteristics of each audio representation during the DirAC analysis
or while determining the DirAC parameters.
The combination of the DirAC metadata occurs after determining 125, 126, 126a for each
35 input format the DirAC parameters, diffuseness, direction as well as the pressure contained
in the transmitted transport channels. The DirAC analysis can estimate the parameters from
an intermediate B-format, obtained by converting the input format as explained previously.
24
Alternatively, DirAC parameters can be advantageously estimated without going through Bformat
but directly from the input format, which might further improve the estimation
accuracy. For example in [7], it is proposed to estimate the diffuseness direct from higher
order Ambisonics. In case of audio objects, a simple metadata convertor 150 in Fig. 15 can
extract from the object metadata direction and diffuseness 5 for each object.
The combination 144 of the several Dirac metadata streams into a single combined DirAC
metadata stream can be achieved as proposed in [4]. For some content it is much better to
10 directly estimate the DirAC parameters from the original format rather than converting it to
a combined B-format first before performing a DirAC analysis. Indeed, the parameters,
direction and diffuseness, can be biased when going to a B-format [3] or when combining
the different sources. Moreover, this alternative allows a
15 Another simpler alternative can average the parameters of the different sources by
weighting them according to their energies:
𝜓(𝑘, 𝑛) =
1
Σ 𝐸􀯜(𝑘, 𝑛) 􀯇􀯜
􀭀􀬵
􀷍𝐸􀯜 (𝑘, 𝑛)
􀯇
􀯜􀭀􀬵
𝜓􀯜(𝑘, 𝑛)
𝑒􀮽􀯈􀮺(𝑘, 𝑛) =
1
Σ (1 − 𝜓􀯜(𝑘, 𝑛))𝐸􀯜(𝑘, 𝑛) 􀯇􀯜
􀭀􀬵
􀷍(1 − 𝜓􀯜(𝑘, 𝑛))𝐸􀯜(𝑘, 𝑛)𝑒􀮽􀯈􀮺
􀯜 (𝑘, 𝑛)
􀯇
􀯜􀭀􀬵
20
For each object there is the possibility to still send its own direction and optionally distance,
diffuseness or any other relevant object attributes as part of the transmitted bitstream from
the encoder to the decoder (see e.g., Figs. 4a, 4b). This extra side-information will enrich
the combined DirAC metadata and will allow the decoder to restitute and or manipulate the
25 object separately. Since an object has a single direction throughout all frequency bands and
can be considered either static or slowly moving, the extra information requires to be
updated less frequently than other DirAC parameters and will engender only very low
additional bit-rate.
30 At the decoder side, directional filtering can be performed as educated in [5] for manipulating
objects. Directional filtering is based upon a short-time spectral attenuation technique. It is
performed in the spectral domain by a zero-phase gain function, which depends upon the
direction of the objects. The direction can be contained in the bitstream if directions of
objects were transmitted as side-information. Otherwise, the direction could also be given
35 interactively by the user.
25
3rd Alternative: combination at decoder side
Alternatively, the combination can be performed at the decoder side. Fig. 14 is a system
overview of the DirAC-based encoder/decoder combining different input 5 formats at decoder
side through a DirAC metadata combiner. In Fig. 14, the DirAC-based coding scheme works
at higher bit rates than previously but allows for the transmission of individual DirAC
metadata. The different DirAC metadata streams are combined 144 as for example
proposed in [4] in the decoder before the DirAC synthesis 220, 240. The DirAC metadata
10 combiner 144 can also obtain the position of an individual object for subsequent
manipulation of the object in DirAC analysis.
Fig. 15 is a system overview of the DirAC-based encoder/decoder combining different input
formats at decoder side in DirAC synthesis. If bit-rate allows, the system can further be
15 enhanced as proposed in Fig. 15 by sending for each input component (FOA/HOA, MC,
Object) its own downmix signal along with its associated DirAC metadata. Still, the different
DirAC streams share a common DirAC synthesis 220, 240 at the decoder to reduce
complexity.
20 Fig. 2a illustrates a concept for performing a synthesis of a plurality of audio scenes in
accordance with a further, second aspect of the present invention. An apparatus illustrated
in Fig. 2a comprises an input interface 100 for receiving a first DirAC description of a first
scene and for receiving a second DirAC description of a second scene and one or more
transport channels.
25
Furthermore, a DirAC synthesizer 220 is provided for synthesizing the plurality of audio
scenes in a spectral domain to obtain a spectral domain audio signal representing the
plurality of audio scenes. Furthermore, a spectrum-time converter 214 is provided that
converts the spectral domain audio signal into a time domain in order to output a time
30 domain audio signal that can be output by speakers, for example. In this case, the DirAC
synthesizer is configured to perform rendering of loudspeaker output signal. Alternatively,
the audio signal could be a stereo signal that can be output to a headphone. Again,
alternatively, the audio signal output by the spectrum-time converter 214 can be a B-format
sound field description. All these signals, i.e., loudspeaker signals for more than two
35 channels, headphone signals or sound field descriptions are time domain signal for further
processing such as outputting by speakers or headphones or for transmission or storage in
the case of sound field descriptions such as first order Ambisonics signals or higher order
Ambisonics signals.
26
Furthermore, the Fig. 2a device additionally comprises a user interface 260 for controlling
the DirAC synthesizer 220 in the spectral domain. Additionally, one or more transport
channels can be provided to the input interface 100 that are to be used together with the
first and second DirAC descriptions that are, in this case, parametric descriptions 5 providing,
for each time/frequency tile, a direction of arrival information and, optionally, additionally a
diffuseness information.
Typically, the two different DirAC descriptions input into the interface 100 in Fig. 2a describe
10 two different audio scenes. In this case, the DirAC synthesizer 220 is configured to perform
a combination of these audio scenes. One alternative of the combination is illustrated in Fig.
2b. Here, a scene combiner 221 is configured to combine the two DirAC description in the
parametric domain, i.e., the parameters are combined to obtain combined direction of arrival
(DoA) parameters and optionally diffuseness parameters at the output of block 221. This
15 data is then introduced into to the DirAC renderer 222 that receives, additionally, the one or
more transport channels in order to channels in order to obtain the spectral domain audio
signal 222. The combination of the DirAC parametric data is preferably performed as
illustrated in Fig. 1d and, as is described with respect to this figure and, particularly, with
respect to the first alternative.
20
Should at least one of the two descriptions input into the scene combiner 221 include
diffuseness values of zero or no diffuseness values at all, then, additionally, the second
alternative can be applied as well as discussed in the context of Fig. 1d.
25 Another alternative is illustrated in Fig. 2c. In this procedure, the individual DirAC
descriptions are rendered by means of a first DirAC renderer 223 for the first description
and a second DirAC renderer 224 for the second description and at the output of blocks
223 and 224, a first and the second spectral domain audio signal are available, and these
first and second spectral domain audio signals are combined within the combiner 225 to
30 obtain, at the output of the combiner 225, a spectral domain combination signal.
Exemplarily, the first DirAC renderer 223 and the second DirAC renderer 224 are configured
to generate a stereo signal having a left channel L and a right channel R. Then, the combiner
225 is configured to combine the left channel from block 223 and the left channel from block
35 224 to obtain a combined left channel. Additionally, the right channel from block 223 is
added with the right channel from block 224, and the result is a combined right channel at
the output of block 225.
27
For individual channels of a multichannel signal, the analogous procedure is performed, i.e.,
the individual channels are individually added, so that always the same channel from a
DirAC renderer 223 is added to the corresponding same channel of the other DirAC
renderer and so on. The same procedure is also performed for, for example, B-format or
higher order Ambisonics signals. When, for example, the first DirAC renderer 5 223 outputs
signals W, X, Y, Z signals, and the second DirAC renderer 224 outputs a similar format,
then the combiner combines the two omnidirectional signals to obtain a combined
omnidirectional signal W, and the same procedure is performed also for the corresponding
components in order to finally obtain a X, Y and a Z combined component.
10
Furthermore, as already outlined with respect to Fig. 2a, the input interface is configured to
receive extra audio object metadata for an audio object. This audio object can already be
included in the first or the second DirAC description or is separate from the first and the
second DirAC description. In this case, the DirAC synthesizer 220 is configured to
15 selectively manipulate the extra audio object metadata or object data related to this extra
audio object metadata to, for example, perform a directional filtering based on the extra
audio object metadata or based on user-given direction information obtained from the user
interface 260. Alternatively or additionally, and as illustrated in Fig. 2d, the DirAC
synthesizer 220 is configured for performing, in the spectral domain, a zero-phase gain
20 function, the zero-phase gain function depending upon a direction of an audio object,
wherein the direction is contained in a bit stream if directions of objects are transmitted as
side information, or wherein the direction of is received from the user interface 260. The
extra audio object metadata input into the interface 100 as an optional feature in Fig. 2a
reflects the possibility to still send, for each individual object its own direction and optionally
25 distance, diffuseness and any other relevant object attributes as part of the transmitted bit
stream from the encoder to the decoder. Thus, the extra audio object metadata may related
to an object already included in the first DirAC description or in the second DirAC description
or is an additional object not included in the first DirAC description and in the second DirAC
description already.
30
However, it is preferred to have the extra audio object metadata already in a DirAC-style,
i.e., a direction of arrival information and, optionally, a diffuseness information although
typical audio objects have a diffusion of zero, i.e., or concentrated to their actual position
resulting in a concentrated and specific direction of arrival that is constant over all frequency
35 bands and that is, with respect to the frame rate, either static or slowly moving. Thus, since
such an object has a single direction throughout all frequency bands and can be considered
either static or slowly moving, the extra information requires to be updated less frequently
than other DirAC parameters and will, therefore, incur only very low additional bitrate.
28
Exemplarily, while the first and the second DirAC descriptions have DoA data and
diffuseness data for each spectral band and for each frame, the extra audio object metadata
only requires a single DoA data for all frequency bands and this data only for every second
frame or, preferably, every third, fourth, fifth or even every tenth frame in the preferred
5 embodiment.
Furthermore, with respect to directional filtering performed in the DirAC synthesizer 220 that
is typically included within a decoder on a decoder side of an encoder/decoder system, the
DirAC synthesizer can, in the Fig. 2b alternative, perform the directional filtering within the
10 parameter domain before the scene combination or again perform the directional filtering
subsequent to the scene combination. However, in this case, the directional filtering is
applied to the combined scene rather than the individual descriptions.
Furthermore, in case an audio object is not included in the first or the second description,
15 but is included by its own audio object metadata, the directional filtering as illustrated by the
selective manipulator can be selectively applied only the extra audio object, for which the
extra audio object metadata exists without effecting the first or the second DirAC description
or the combined DirAC description. For the audio object itself, there either exists a separate
transport channel representing the object waveform signal or the object waveforms signal
20 is included in the downmixed transport channel.
A selective manipulation as illustrated, for example, in Fig. 2b may, for example, proceed in
such a way that a certain direction of arrival is given by the direction of audio object
introduced in Fig. 2d included in the bit stream as side information or received from a user
25 interface. Then, based on the user-given direction or control information, the user may, for
example, outline that, from a certain direction, the audio data is to be enhanced or is to be
attenuated. Thus, the object (metadata) for the object under consideration is amplified or
attenuated.
30 In the case of actual waveform data as the object data introduced into the selective
manipulator 226 from the left in Fig. 2d, the audio data would be actually attenuated or
enhanced depending on the control information. However, in the case of object data having,
in addition to direction of arrival and optionally diffuseness or distance, a further energy
information, then the energy information for the object would be reduced in the case of a
35 required attenuation for the object or the energy information would be increased in the case
of a required amplification of the object data.
29
Thus, the directional filtering is based upon a short-time spectral attenuation technique, and
it is performed it the spectral domain by a zero-phase gain function which depends upon
the direction of the objects. The direction can be contained in the bit stream if directions of
objects were transmitted as side-information. Otherwise, the direction could also be given
interactively by the user. Naturally, the same procedure cannot only 5 be applied to the
individual object given and reflected by the extra audio object metadata typically provided
by DoA data for all frequency bands and DoA data with a low update ratio with respect to
the frame rate and also given by the energy information for the object, but the directional
filtering can also be applied to the first DirAC description independent from the second
10 DirAC description or vice versa or can be also applied to the combined DirAC description
as the case may be.
Furthermore, it is to be noted that the feature with respect to the extra audio object data can
also be applied in the first aspect of the present invention illustrated with respect to Figs. 1a
15 to 1f. Then, the input interface 100 of Fig. 1a additionally receives the extra audio object
data as discussed with respect to Fig. 2a, and the format combiner may be implemented as
the DirAC synthesizer in the spectral domain 220 controlled by a user interface 260.
Furthermore, the second aspect of the present invention as illustrated in Fig. 2 is different
20 from the first aspect in that the input interface receives already two DirAC descriptions, i.e.,
descriptions of a sound field that are in the same format and, therefore, for the second
aspect, the format converter 120 of the first aspect is not necessarily required.
On the other hand, when the input into the format combiner 140 of Fig. 1a consists of two
25 DirAC descriptions, then the format combiner 140 can be implemented as discussed with
respect to the second aspect illustrated in Fig. 2a, or, alternatively, the Fig. 2a devices 220,
240, can be implemented as discussed with respect to the format combiner 140 of Fig. 1a
of the first aspect.
30 Fig. 3a illustrates an audio data converter comprising an input interface 100 for receiving
an object description of an audio object having audio object metadata. Furthermore, the
input interface 100 is followed by a metadata converter 150 also corresponding to the
metadata converters 125, 126 discussed with respect to the first aspect of the present
invention for converting the audio object metadata into DirAC metadata. The output of the
35 Fig. 3a audio converter is constituted by an output interface 300 for transmitting or storing
the DirAC metadata. The input interface 100 may, additionally receive a waveform signal
as illustrated by the second arrow input into the interface 100. Furthermore, the output
interface 300 may be implemented to introduce, typically an encoded representation of the
30
waveform signal into the output signal output by block 300. If the audio data converter is
configured to only convert a single object description including metadata, then the output
interface 300 also provides a DirAC description of this single audio object together with the
typically encoded waveform signal as the DirAC transport channel.
5
Particularly, the audio object metadata has an object position, and the DirAC metadata has
a direction of arrival with respect to a reference position derived from the object position.
Particularly, the metadata converter 150, 125, 126 is configured to convert DirAC
parameters derived from the object data format into pressure/velocity data, and the
10 metadata converter is configured to apply a DirAC analysis to this pressure/velocity data
as, for example, illustrated by the flowchart of Fig. 3c consisting of block 302, 304, 306. For
this purpose, the DirAC parameters output by block 306 have a better quality than the DirAC
parameters derived from the object metadata obtained by block 302, i.e., are enhanced
DirAC parameters. Fig. 3b illustrates the conversion of a position for an object into the
15 direction of arrival with respect to a reference position for the specific object.
Fig. 3f illustrates a schematic diagram for explaining the functionality of the metadata
converter 150. The metadata converter 150 receives the position of the object indicated by
vector P in a coordinate system. Furthermore, the reference position, to which the DirAC
20 metadata are to be related is given by vector R in the same coordinate system. Thus, the
direction of arrival vector DoA extends from the tip of vector R to the tip of vector B. Thus,
the actual DoA vector is obtained by subtracting the reference position R vector from the
object position P vector.
25 In order to have a normalized DoA information indicated by the vector DoA, the vector
difference is divided by the magnitude or length of the vector DoA. Furthermore, and should
this be necessary and intended, the length of the DoA vector can also be included into the
metadata generated by the metadata converter 150 so that, additionally, the distance of the
object from the reference point is also included in the metadata so that a selective
30 manipulation of this object can also be performed based on the distance of the object from
the reference position. Particularly, the extract direction block 148 of Fig. 1f may also
operate as discussed with respect to Fig. 3f, although other alternatives for calculating the
DoA information and, optionally, the distance information can be applied as well.
Furthermore, as already discussed with respect to Fig. 3a, blocks 125 and 126 illustrated in
35 Fig. 1c or 1d may operate in the similar way as discussed with respect to Fig. 3f.
Furthermore, the Fig. 3a device may be configured to receive a plurality of audio object
descriptions, and the metadata converter is configured to convert each metadata
31
description directly into a DirAC description and, then, the metadata converter is configured
to combine the individual DirAC metadata descriptions to obtain a combined DirAC
description as the DirAC metadata illustrated in Fig. 3a. In one embodiment, the
combination is performed by calculating 320 a weighting factor for a first direction of arrival
using a first energy and by calculating 322 a weighting factor for a second direction 5 of arrival
using a second energy, where the direction of arrival is processed by blocks 320, 332
related to the same time/frequency bin. Then, in block 324, a weighted addition is performed
as also discussed with respect to item 144 in Fig. 1d. Thus, the procedure illustrated in Fig.
3a represents an embodiment of the first alternative Fig. 1d.
10
However, with respect to the second alternative, the procedure would be that all diffuseness
are set to zero or to a small value and, for a time/frequency bin, all different direction of
arrival values that are given for this time/frequency bin are considered and the largest
direction of arrival value is selected to be the combined direction of arrival value for this
15 time/frequency bin. In other embodiments, one could also select the second to largest value
provided that the energy information for these two direction of arrival values are not so
different. The direction of arrival value is selected whose energy is either the largest energy
among the energies from the different contribution for this time frequency bin or the second
or the third highest energy.
20
Thus, the third aspect as described with respect to Figs. 3a to 3f are different from the first
aspect in that the third aspect is also useful for the conversion of a single object description
into a DirAC metadata. Alternatively, the input interface 100 may receive several object
descriptions that are in the same object/metadata format. Thus, any format converter as
25 discussed with respect to the first aspect in Fig. 1a is not required. Thus, the Fig. 3a
embodiment may be useful in the context of receiving two different object descriptions using
different object waveform signals and different object metadata as the first scene description
and the second description as input into the format combiner 140, and the output of the
metadata converter 150, 125, 126 or 148 may be a DirAC representation with DirAC
30 metadata and, therefore, the DirAC analyzer 180 of Fig. 1 is also not required. However,
the other elements with respect to the transport channel generator 160 corresponding to
the downmixer 163 of Fig. 3a can be used in the context of the third aspect as well as the
transport channel encoder 170, the metadata encoder 190 and, in this context, the output
interface 300 of Fig. 3a corresponds to the output interface 200 of Fig. 1a. Hence, all
35 corresponding descriptions given with respect to the first aspect also apply to the third
aspect as well.
32
Figs. 4a, 4b illustrate a fourth aspect of the present invention in the context of an apparatus
for performing a synthesis of audio data. Particularly, the apparatus has an input interface
100 for receiving a DirAC description of an audio scene having DirAC metadata and
additionally for receiving an object signal having object metadata. This audio scene encoder
illustrated in Fig. 4b additionally comprises the metadata generator 400 5 for generating a
combined metadata description comprising the DirAC metadata on the one hand and the
object metadata on the other hand. The DirAC metadata comprises the direction of arrival
for individual time/frequency tiles and the object metadata comprises a direction or
additionally a distance or a diffuseness of an individual object.
10
Particularly, the input interface 100 is configured to receive, additionally, a transport signal
associated with the DirAC description of the audio scene as illustrated in Fig. 4b, and the
input interface is additionally configured for receiving an object waveform signal associated
with the object signal. Therefore, the scene encoder further comprises a transport signal
15 encoder for encoding the transport signal and the object waveform signal, and the transport
encoder 170 may correspond to the encoder 170 of Fig. 1a.
Particularly, the metadata generator 140 that generates the combined metadata may be
configured as discussed with respect to the first aspect, the second aspect or the third
20 aspect. And, in a preferred embodiment, the metadata generator 400 is configured to
generate, for the object metadata, a single broadband direction per time, i.e., for a certain
time frame, and the metadata generator is configured to refresh the single broadband
direction per time less frequently than the DirAC metadata.
25 The procedure discussed with respect to Fig. 4b allows to have combined metadata that
has metadata for a full DirAC description and that has, in addition, metadata for an additional
audio object, but in the DirAC format so that a very useful DirAC rendering can be performed
by, at the same time, a selective directional filtering or modification as already discussed
with respect to the second aspect can be performed.
30
Thus, the fourth aspect of the present invention and, particularly, the metadata generator
400 represents a specific format converter where the common format is the DirAC format,
and the input is a DirAC description for the first scene in the first format discussed with
respect to Fig. 1a and the second scene is a single or a combined such as SAOC object
35 signal. Hence, the output of the format converter 120 represents the output of the metadata
generator 400 but, in contrast to an actual specific combination of the metadata by one of
the two alternatives, for example, as discussed with respect to Fig. 1d, the object metadata
33
is included in the output signal, i.e., the “combined metadata” separate from the metadata
for the DirAC description to allow a selective modification for the object data.
Thus, the “direction/distance/diffuseness” indicated at item 2 at the right hand side of Fig.
4a corresponds to the extra audio object metadata input into the input interface 5 100 of Fig.
2a, but, in the embodiment of Fig. 4a, for a single DirAC description only. Thus, in a sense,
one could say that Fig. 2a represents a decoder-side implementation of the encoder
illustrated in Fig. 4a, 4b with the provision that the decoder side of Fig. 2a device receives
only a single DirAC description and the object metadata generated by the metadata
10 generator 400 within the same bit stream as the “extra audio object metadata”.
Thus, a completely different modification of the extra object data can be performed when
the encoded transport signal has a separate representation of the object waveform signal
separate from the DirAC transport stream. And, however, the transport encoder 170
15 downmixes both data, i.e., the transport channel for the DirAC description and the waveform
signal from the object, then the separation will be less perfect, but by means of additional
object energy information, even a separation from a combined downmix channel and a
selective modification of the object with respect to the DirAC description is available.
20 Fig. 5a to 5d represent a further of fifth aspect of the invention in the context of an apparatus
for performing a synthesis of audio data. To this end, an input interface 100 is provided for
receiving a DirAC description of one or more audio objects and/or a DirAC description of a
multi-channel signal and/or a DirAC description of a first order Ambisonics signal and/or a
higher order Ambisonics signal, wherein the DirAC description comprises position
25 information of the one or more objects or a side information for the first order Ambisonics
signals or the high order Ambisonics signals or a position information for the multi-channel
signal as side information or from a user interface.
Particularly, a manipulator 500 is configured for manipulating the DirAC description of the
30 one or more audio objects, the DirAC description of the multi-channel signal, the DirAC
description of the first order Ambisonics signals or the DirAC description of the high order
Ambisonics signals to obtain a manipulated DirAC description. In order to synthesize this
manipulated DirAC description, a DirAC synthesizer 220, 240 is configured for synthesizing
this manipulated DirAC description to obtain synthesized audio data.
35
In a preferred embodiment, the DirAC synthesizer 220, 240 comprises a DirAC renderer
222 as illustrated in Fig. 5b and the subsequently connected spectral-time converter 240
34
that outputs the manipulated time domain signal. Particularly, the manipulator 500 is
configured to perform a position-dependent weighting operation prior to DirAC rendering.
Particularly, when the DirAC synthesizer is configured to output a plurality of objects of a
first order Ambisonics signals or a high order Ambisonics signal or a multi-5 channel signal,
the DirAC synthesizer is configured to use a separate spectral-time converter for each
object or each component of the first or the high order Ambisonics signals or for each
channel of the multichannel signal as illustrated in Fig. 5d at blocks 506, 508. As outlined in
block 510 then the output of the corresponding separate conversions are added together
10 provided that all the signals are in a common format, i.e., in compatible format.
Therefore, in case of the input interface 100 of Fig. 5a, receiving more than one, i.e., two or
three representations, each representation could be manipulated separately as illustrated
in block 502 in the parameter domain as already discussed with respect to Fig. 2b or 2c,
15 and, then, a synthesis could be performed as outlined in block 504 for each manipulated
description, and the synthesis could then be added in the time domain as discussed with
respect to block 510 in Fig. 5d. Alternatively, the result of the individual DirAC synthesis
procedures in the spectral domain could already be added in the spectral domain and then
a single time domain conversion could be used as well. Particularly, the manipulator 500
20 may be implemented as the manipulator discussed with respect to Fig. 2d or discussed with
respect to any other aspect before.
Hence, the fifth aspect of the present invention provides a significant feature with respect
to the fact, when individual DirAC descriptions of very different sound signals are input, and
25 when a certain manipulation of the individual descriptions is performed as discussed with
respect to block 500 of Fig. 5a, where an input into the manipulator 500 may be a DirAC
description of any format, including only a single format, while the second aspect was
concentrating on the reception of at least two different DirAC descriptions or where the
fourth aspect, for example, was related to the reception of a DirAC description on the one
30 hand and an object signal description on the other hand.
Subsequently, reference is made to Fig. 6. Fig. 6 illustrates another implementation for
performing a synthesis different from the DirAC synthesizer. When, for example, a sound
field analyzer generates, for each source signal, a separate mono signal S and an original
35 direction of arrival and when, depending on the translation information, a new direction of
arrival is calculated, then the Ambisonics signal generator 430 of Fig. 6, for example, would
be used to generate a sound field description for the sound source signal, i.e., the mono
signal S but for the new direction of arrival (DoA) data consisting of a horizontal angle θ or
35
an elevation angle θ and an azimuth angle φ. Then, a procedure performed by the sound
field calculator 420 of Fig. 6 would be to generate, for example, a first-order Ambisonics
sound field representation for each sound source with the new direction of arrival and, then,
a further modification per sound source could be performed using a scaling factor
depending on the distance of the sound field to the new reference location 5 and, then, all the
sound fields from the individual sources could superposed to each other to finally obtain the
modified sound field, once again, in, for example, an Ambisonics representation related to
a certain new reference location.
10 When one interprets that each time/frequency bin processed by the DirAC analyzer 422
represents a certain (bandwidth limited) sound source, then the Ambisonics signal
generator 430 could be used, instead of the DirAC synthesizer 425, to generate, for each
time/frequency bin, a full Ambisonics representation using the downmix signal or pressure
signal or omnidirectional component for this time/frequency bin as the “mono signal S” of
15 Fig. 6. Then, an individual frequency-time conversion in frequency-time converter 426 for
each of the W, X, Y, Z component would then result in a sound field description different
from what is illustrated in Fig. 6.
Subsequently, further explanations regarding a DirAC analysis and a DirAC synthesis are
20 given as known in the art. Fig. 7a illustrates a DirAC analyzer as originally disclosed, for
example, in the reference “Directional Audio Coding” from IWPASH of 2009. The DirAC
analyzer comprises a bank of band filters 1310, an energy analyzer 1320, an intensity
analyzer 1330, a temporal averaging block 1340 and a diffuseness calculator 1350 and the
direction calculator 1360. In DirAC, both analysis and synthesis are performed in the
25 frequency domain. There are several methods for dividing the sound into frequency bands,
within distinct properties each. The most commonly used frequency transforms include
short time Fourier transform (STFT), and Quadrature mirror filter bank (QMF). In addition to
these, there is a full liberty to design a filter bank with arbitrary filters that are optimized to
any specific purposes. The target of directional analysis is to estimate at each frequency
30 band the direction of arrival of sound, together with an estimate if the sound is arriving from
one or multiple directions at the same time. In principle, this can be performed with a number
of techniques, however, the energetic analysis of sound field has been found to be suitable,
which is illustrated in Fig. 7a. The energetic analysis can be performed, when the pressure
signal and velocity signals in one, two or three dimensions are captured from a single
35 position. In first-order B-format signals, the omnidirectional signal is called W-signal, which
36
has been scaled down by the square root of two. The sound pressure can be estimated as
𝑆 = √2 ∗ 𝑊, expressed in the STFT domain.
The X-, Y- and Z channels have the directional pattern of a dipole directed along the
Cartesian axis, which form together a vector U = [X, Y, Z]. The vector estimates 5 the sound
field velocity vector, and is also expressed in STFT domain. The energy E of the sound field
is computed. The capturing of B-format signals can be obtained with either coincident
positioning of directional microphones, or with a closely-spaced set of omnidirectional
microphones. In some applications, the microphone signals may be formed in a
10 computational domain, i.e., simulated. The direction of sound is defined to be the opposite
direction of the intensity vector I. The direction is denoted as corresponding angular azimuth
and elevation values in the transmitted metadata. The diffuseness of sound field is also
computed using an expectation operator of the intensity vector and the energy. The
outcome of this equation is a real-valued number between zero and one, characterizing if
15 the sound energy is arriving from a single direction (diffuseness is zero), or from all
directions (diffuseness is one). This procedure is appropriate in the case when the full 3D
or less dimensional velocity information is available.
Fig. 7b illustrates a DirAC synthesis, once again having a bank of band filters 1370, a virtual
20 microphone block 1400, a direct/diffuse synthesizer block 1450, and a certain loudspeaker
setup or a virtual intended loudspeaker setup 1460. Additionally, a diffuseness-gain
transformer 1380, a vector based amplitude panning (VBAP) gain table block 1390, a
microphone compensation block 1420, a loudspeaker gain averaging block 1430 and a
distributer 1440 for other channels is used. In this DirAC synthesis with loudspeakers, the
25 high quality version of DirAC synthesis shown in Fig. 7b receives all B-format signals, for
which a virtual microphone signal is computed for each loudspeaker direction of the
loudspeaker setup 1460. The utilized directional pattern is typically a dipole. The virtual
microphone signals are then modified in non-linear fashion, depending on the metadata.
The low bitrate version of DirAC is not shown in Fig. 7b, however, in this situation, only one
30 channel of audio is transmitted as illustrated in Fig. 6. The difference in processing is that
all virtual microphone signals would be replaced by the single channel of audio received.
The virtual microphone signals are divided into two streams: the diffuse and the non-diffuse
streams, which are processed separately.
35 The non-diffuse sound is reproduced as point sources by using vector base amplitude
panning (VBAP). In panning, a monophonic sound signal is applied to a subset of
37
loudspeakers after multiplication with loudspeaker-specific gain factors. The gain factors
are computed using the information of a loudspeaker setup, and specified panning direction.
In the low-bit-rate version, the input signal is simply panned to the directions implied by the
metadata. In the high-quality version, each virtual microphone signal is multiplied with the
corresponding gain factor, which produces the same effect with panning, 5 however it is less
prone to any non-linear artifacts.
In many cases, the directional metadata is subject to abrupt temporal changes. To avoid
artifacts, the gain factors for loudspeakers computed with VBAP are smoothed by temporal
10 integration with frequency-dependent time constants equaling to about 50 cycle periods at
each band. This effectively removes the artifacts, however, the changes in direction are not
perceived to be slower than without averaging in most of the cases. The aim of the synthesis
of the diffuse sound is to create perception of sound that surrounds the listener. In the lowbit-
rate version, the diffuse stream is reproduced by decorrelating the input signal and
15 reproducing it from every loudspeaker. In the high-quality version, the virtual microphone
signals of diffuse stream are already incoherent in some degree, and they need to be
decorrelated only mildly. This approach provides better spatial quality for surround
reverberation and ambient sound than the low bit-rate version. For the DirAC synthesis with
headphones, DirAC is formulated with a certain amount of virtual loudspeakers around the
20 listener for the non-diffuse stream and a certain number of loudspeakers for the diffuse
steam. The virtual loudspeakers are implemented as convolution of input signals with a
measured head-related transfer functions (HRTFs).
Subsequently, a further general relation with respect to the different aspects and,
25 particularly, with respect to further implementations of the first aspect as discussed with
respect to Fig. 1a is given. Generally, the present invention refers to the combination of
different scenes in different formats using a common format, where the common format
may, for example, be the B-format domain, the pressure/velocity domain or the metadata
domain as discussed, for example, in items 120, 140 of Fig. 1a.
30
When the combination is not done directly in the DirAC common format, then a DirAC
analysis 802 is performed in one alternative before the transmission in the encoder as
discussed before with respect to item 180 of Fig. 1a.
35 Then, subsequent to the DirAC analysis, the result is encoded as discussed before with
respect to the encoder 170 and the metadata encoder 190 and the encoded result is
38
transmitted via the encoded output signal generated by the output interface 200. However,
in a further alternative, the result could be directly rendered by a Fig. 1a device when the
output of block 160 of Fig. 1a and the output of block 180 of Fig. 1a is forwarded to a DirAC
renderer. Thus, the Fig. 1a device would not be a specific encoder device but would be an
analyzer and a corresponding 5 renderer.
A further alternative is illustrated in the right branch of Fig. 8, where a transmission from the
encoder to the decoder is performed and, as illustrated in block 804, the DirAC analysis and
the DirAC synthesis are performed subsequent to the transmission, i.e., at a decoder-side.
10 This procedure would be the case, when the alternative of Fig. 1a is used, i.e., that the
encoded output signal is a B-format signal without spatial metadata. Subsequent to block
808, the result could be rendered for replay or, alternatively, the result could even be
encoded and again transmitted. Thus, it becomes clear that the inventive procedures as
defined and described with respect to the different aspects are highly flexible and can be
15 very well adapted to specific use cases.
1st Aspect of Invention: Universal DirAC-based spatial audio coding/rendering
A Dirac-based spatial audio coder that can encode multi-channel signals, Ambisonics
20 formats and audio objects separately or simultaneously.
Benefits and Advantages over State of the Art
• Universal DirAC-based spatial audio coding scheme for the most relevant immersive
25 audio input formats
• Universal audio rendering of different input formats on different output formats
2nd Aspect of Invention: Combining two or more DirAC descriptions on a decoder
30 The second aspect of the invention is related to the combination and rendering two or more
DirAC descriptions in the spectral domain.
Benefits and Advantages over State of the Art
35 • Efficient and precise DirAC stream combination
• Allows the usage of DirAC universally represent any scene and to efficiently
combine different streams in the parameter domain or the spectral domain
39
• Efficient and intuitive scene manipulation of individual DirAC scenes or the
combined scene in the spectral domain and subsequent conversion into the time
domain of the manipulated combined scene.
3rd Aspect of Invention: Conversion of audio objects into the DirAC domain
5
The third aspect of the invention is related to the conversion of object metadata and
optionally object waveform signals directly into the DirAC domain and in an embodiment the
combination of several objects into an object representation.
10 Benefits and Advantages over State of the Art
• Efficient and precise DirAC metadata estimation by simple metadata transcoder of
the audio objects metadata
• Allows DirAC to code complex audio scenes involving one or more audio objects
15 • Efficient method for coding audio objects through DirAC in a single parametric
representation of the complete audio scene.
4th Aspect of Invention: Combination of Object metadata and regular DirAC metadata
The third aspect of the invention addresses the amendment of the DirAC metadata with the
20 directions and, optimally, the distance or diffuseness of the individual objects composing
the combined audio scene represented by the DirAC parameters. This extra information is
easily coded, since it consist mainly of a single broadband direction per time unit and can
be refreshed less frequently than the other DirAC parameters since objects can be assumed
to be either static or moving at a slow pace.
25
Benefits and Advantages over State of the Art
• Allows DirAC to code a complex audio scene involving one or more audio objects
• Efficient and precise DirAC metadata estimation by simple metadata transcoder of
30 the audio objects metadata.
• More efficient method for coding audio objects through DirAC by combining
efficiently their metadata in DirAC domain
• Efficient method for coding audio objects and through DirAC by combining efficiently
their audio representations in a single parametric representation of the audio scene.
40
5th Aspect of Invention: Manipulation of Objects MC scenes and FOA/HOA C in DirAC
synthesis
The fourth aspect is related to the decoder side and exploits the known 5 positions of audio
objects. The positions can be given by the user though an interactive interface and can also
be included as extra side-information within the bitstream.
The aim is to be able to manipulate an output audio scene comprising a number of objects
10 by individually changing the objects’ attributes such as levels, equalization and/or spatial
positions. It can also be envisioned to filter completely the object or restitute individual
objects from the combined stream.
The manipulation of the output audio scene can be achieved by jointly processing the spatial
15 parameters of the DirAC metadata, the objects’ metadata, interactive user input if present
and the audio signals carried in the transport channels.
Benefits and Advantages over State of the Art
20 • Allows DirAC to output at the decoder side audio objects as presented at the input
of the encoder.
• Allows DirAC reproduction to manipulate individual audio object by applying gains,
rotation , or…
• Capability requires minimal additional computational effort since it only requires a
25 position-dependent weighting operation prior to the rendering & synthesis filterbank
at the end of the DirAC synthesis (additional object outputs will just require one
additional synthesis filterbank per object output).
References that are all incorporated it their entirety by reference:
30
[1] V. Pulkki, M-V Laitinen, J Vilkamo, J Ahonen, T Lokki and T Pihlajamäki, “Directional
audio coding - perception-based reproduction of spatial sound”, International Workshop on
the Principles and Application on Spatial Hearing, Nov. 2009, Zao; Miyagi, Japan.
35 [2] Ville Pulkki. “Virtual source positioning using vector base amplitude panning”. J. Audio
Eng. Soc., 45(6):456{466, June 1997.
41
[3] M. V. Laitinen and V. Pulkki, "Converting 5.1 audio recordings to B-format for directional
audio coding reproduction," 2011 IEEE International Conference on Acoustics, Speech and
Signal Processing (ICASSP), Prague, 2011, pp. 61-64.
[4] G. Del Galdo, F. Kuech, M. Kallinger and R. Schultz-Amling, "Efficient 5 merging of
multiple audio streams for spatial sound reproduction in Directional Audio Coding," 2009
IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, 2009,
pp. 265-268.
10 [5] Jürgen HERRE, CORNELIA FALCH, DIRK MAHNE, GIOVANNI DEL GALDO, MARKUS
KALLINGER, AND OLIVER THIERGART, “Interactive Teleconferencing Combining Spatial
Audio Object Coding and DirAC Technology”, J. Audio Eng. Soc., Vol. 59, No. 12, 2011
December.
15 [6] R. Schultz-Amling, F. Kuech, M. Kallinger, G. Del Galdo, J. Ahonen, V. Pulkki, “Planar
Microphone Array Processing for the Analysis and Reproduction of Spatial Audio using
Directional Audio Coding,” Audio Engineering Society Convention 124, Amsterdam, The
Netherlands, 2008.
20 [7] Daniel P. Jarrett and Oliver Thiergart and Emanuel A. P. Habets and Patrick A. Naylor,
“Coherence-Based Diffuseness Estimation in the Spherical Harmonic Domain”, IEEE 27th
Convention of Electrical and Electronics Engineers in Israel (IEEEI), 2012.
[8] US Patent 9,015,051.
25
The present invention provides, in further embodiments, and particularly with respect to the
first aspect and also with respect to the other aspects different alternatives. These
alternatives are the following:
30 Firstly, combining different formats in the B format domain and either doing the DirAC
analysis in the encoder or transmitting the combined channels to a decoder and doing the
DirAC analysis and synthesis there.
Secondly, combining different formats in the pressure/velocity domain and doing the DirAC
35 analysis in the encoder. Alternatively, the pressure/velocity data are transmitted to the
decoder and the DirAC analysis is done in the decoder and the synthesis is also done in
the decoder.
42
Thirdly, combining different formats in the metadata domain and transmitting a single DirAC
stream or transmitting several DirAC streams to a decoder before combining them and
doing the combination in the decoder.
Furthermore, embodiments or aspects of the present invention are related 5 to the following
aspects:
Firstly, combining of different audio formats in accordance with the above three alternatives.
10 Secondly, a reception, combination and rendering of two DirAC descriptions already in the
same format is performed.
Thirdly, a specific object to DirAC converter with a “direct conversion” of object data to DirAC
data is implemented.
15
Fourthly, object metadata in addition to normal DirAC metadata and a combination of both
metadata; both data are existing in the bitstream side-by-side, but audio objects are also
described by DirAC metadata-style.
20 Fifthly, objects and the DirAC stream are separately transmitted to a decoder and objects
are selectively manipulated within the decoder before converting the output audio
(loudspeaker) signals into the time-domain.
It is to be mentioned here that all alternatives or aspects as discussed before and all aspects
25 as defined by independent claims in the following claims can be used individually, i.e.,
without any other alternative or object than the contemplated alternative, object or
independent claim. However, in other embodiments, two or more of the alternatives or the
aspects or the independent claims can be combined with each other and, in other
embodiments, all aspects, or alternatives and all independent claims can be combined to
30 each other.
An inventively encoded audio signal can be stored on a digital storage medium or a nontransitory
storage medium or can be transmitted on a transmission medium such as a
wireless transmission medium or a wired transmission medium such as the Internet.
35
Although some aspects have been described in the context of an apparatus, it is clear that
these aspects also represent a description of the corresponding method, where a block or
device corresponds to a method step or a feature of a method step. Analogously, aspects
43
described in the context of a method step also represent a description of a corresponding
block or item or feature of a corresponding apparatus.
Depending on certain implementation requirements, embodiments of the invention can be
implemented in hardware or in software. The implementation can be 5 performed using a
digital storage medium, for example a floppy disk, a DVD, a CD, a ROM, a PROM, an
EPROM, an EEPROM or a FLASH memory, having electronically readable control signals
stored thereon, which cooperate (or are capable of cooperating) with a programmable
computer system such that the respective method is performed.
10
Some embodiments according to the invention comprise a data carrier having electronically
readable control signals, which are capable of cooperating with a programmable computer
system, such that one of the methods described herein is performed.
15 Generally, embodiments of the present invention can be implemented as a computer
program product with a program code, the program code being operative for performing
one of the methods when the computer program product runs on a computer. The program
code may for example be stored on a machine readable carrier.
20 Other embodiments comprise the computer program for performing one of the methods
described herein, stored on a machine readable carrier or a non-transitory storage medium.
In other words, an embodiment of the inventive method is, therefore, a computer program
having a program code for performing one of the methods described herein, when the
25 computer program runs on a computer.
A further embodiment of the inventive methods is, therefore, a data carrier (or a digital
storage medium, or a computer-readable medium) comprising, recorded thereon, the
computer program for performing one of the methods described herein.
30
A further embodiment of the inventive method is, therefore, a data stream or a sequence of
signals representing the computer program for performing one of the methods described
herein. The data stream or the sequence of signals may for example be configured to be
transferred via a data communication connection, for example via the Internet.
35
A further embodiment comprises a processing means, for example a computer, or a
programmable logic device, configured to or adapted to perform one of the methods
described herein.
44
A further embodiment comprises a computer having installed thereon the computer program
for performing one of the methods described herein.
In some embodiments, a programmable logic device (for example a 5 field programmable
gate array) may be used to perform some or all of the functionalities of the methods
described herein. In some embodiments, a field programmable gate array may cooperate
with a microprocessor in order to perform one of the methods described herein. Generally,
the methods are preferably performed by any hardware apparatus.
10
In the following additional embodiments are described. An embodiment according to the
invention provides an apparatus for generating a description of a combined audio scene,
comprising, an input interface (100) for receiving a first description of a first scene in a first
format and a second description of a second scene in a second format, wherein the second
15 format is different from the first format, a format converter (120) for converting the first
description into a common format and for converting the second description into the
common format, when the second format is different from the common format, and a format
combiner (140) for combining the first description in the common format and the second
description in the common format to obtain the combined audio scene.
20
A further embodiment according to the invention provides apparatus, wherein the first
format and the second format are selected from a group of formats comprising a first order
Ambisonics format, a high order Ambisonics format, the common format, a DirAC format,
an audio object format and a multi-channel format.
25
A further embodiment according to the invention provides apparatus, wherein the format
converter (120) is configured to convert the first description into a first B-format signal
representation and to convert the second description into a second B-format signal
representation, and wherein the format combiner (140) is configured to combine the first
30 and the second B-format signal representation by individually combining the individual
components of the first and the second B-format signal representation.
A further embodiment according to the invention provides apparatus, wherein the format
converter (120) is configured to convert the first description into a first pressure/velocity
35 signal representation and to convert the second description into a second pressure/velocity
signal representation, and wherein the format combiner (140) is configured to combine the
first and the second pressure/velocity signal representation by individually combining the
45
individual components of the pressure/velocity signal representations to obtain a combined
pressure/velocity signal representation.
A further embodiment according to the invention provides apparatus, wherein the format
converter (120) is configured to convert the first description into a first 5 DirAC parameter
representation and to convert the second description into a second DirAC parameter
representation, when the second description is different from the DirAC parameter
representation, and wherein the format combiner (140) is configured to combine the first
and the second DirAC parameter representations by individually combining the individual
10 components of the first and second DirAC parameter representations to obtain a combined
DirAC parameter representation for the combined audio scene.
A further embodiment according to the invention provides apparatus, wherein the format
combiner (140) is configured to generate direction of arrival values for time-frequency tiles
15 or direction of arrival values and diffuseness values for the time-frequency tiles representing
the combined audio scene.
A further embodiment according to the invention provides apparatus, further comprising a
DirAC analyzer (180) for analyzing the combined audio scene to derive DirAC parameters
20 for the combined audio scene, wherein the DirAC parameters comprise direction of arrival
values for time-frequency tiles or direction of arrival values and diffuseness values for the
time-frequency tiles representing the combined audio scene.
Another embodiment according to the invention provides apparatus, further comprising a
25 transport channel generator (160) for generating a transport channel signal from the
combined audio scene or from the first scene and the second scene, and a transport
channel encoder (170) for core encoding the transport channel signal, or wherein the
transport channel generator (160) is configured to generate a stereo signal from the first
scene or the second scene being in a first order Ambisonics or a higher order Ambisonics
30 format using a beam former being directed to a left position or the right position,
respectively, or wherein the transport channel generator (160) is configured to generate a
stereo signal from the first scene or the second scene being in a multichannel representation
by downmixing three or more channels of the multichannel representation, or wherein the
transport channel generator (160) is configured to generate a stereo signal from the first
35 scene or the second scene being in an audio object representation by panning each object
using a position of the object or by downmixing objects into a stereo downmix using
information indicating, which object is located in which stereo channel, or wherein the
transport channel generator (160) is configured to add only the left channel of the stereo
46
signal to the left downmix transport channel and to add only the right channel of the stereo
signal to obtain a right transport channel, or
wherein the common format is the B-format, and wherein the transport channel generator
(160) is configured to process a combined B-format representation to derive the transport
channel signal, wherein the processing comprises performing a beamforming 5 operation or
extracting a subset of components of the B-format signal such as the omnidirectional
component as the mono transport channel, or wherein the processing comprises
beamforming using the omnidirectional signal and the Y component with opposite signs of
the B-format to calculate left and right channels, or wherein the processing comprises a
10 beamforming operation using the components of the B-format and the given azimuth angle
and the given elevation angle, or wherein the transport channel generator (160) is
configured to prove the B-format signals of the combined audio scene to the transport
channel encoder, wherein any spatial metadata are not included in the combined audio
scene output by the format combiner (140).
15
A further embodiment according to the invention provides apparatus, further comprising,
a metadata encoder (190) for encoding DirAC metadata described in the combined audio
scene to obtain encoded DirAC metadata, or for encoding DirAC metadata derived from the
first scene to obtain first encoded DirAC metadata and for encoding DirAC metadata derived
20 from the second scene to obtain second encoded DirAC metadata.
A further embodiment according to the invention provides apparatus, further comprising an
output interface (200) for generating an encoded output signal representing the combined
audio scene, the output signal comprising encoded DirAC metadata and one or more
25 encoded transport channels.
A further embodiment according to the invention provides apparatus, wherein the format
converter (120) is configured to convert a high order Ambisonics or a first order Ambisonics
format into the B-format, wherein the high order Ambisonics format is truncated before being
30 converted into the B-format, or wherein the format converter (120) is configured to project
an object or a channel on spherical harmonics at a reference position to obtain projected
signals, and wherein the format combiner (140) is configured to combine the projection
signals to obtain B-format coefficients, wherein the object or the channel is located in space
at a specified position and has an optional individual distance from a reference position, or
35 wherein the format converter (120) is configured to perform a DirAC analysis comprising a
time-frequency analysis of B-format components and a determination of pressure and
velocity vectors, and wherein the format combiner (140) is configured to combine different
pressure/velocity vectors and wherein the format combiner (140) further comprises a DirAC
47
analyzer for deriving DirAC metadata from the combined pressure/velocity data, or wherein
the format converter (120) is configured to extract DirAC parameters from object metadata
of an audio object format as the first or second format, wherein the pressure vector is the
object waveform signal and the direction is derived from the object position in space or the
diffuseness is directly given in the object metadata or is set to a default 5 value such as 0
value, or wherein the format converter (120) is configured to convert DirAC parameters
derived from the object data format into pressure/velocity data and the format combiner
(140) is configured to combine the pressure/velocity data with pressure/velocity data
derived from a different description of one or more different audio objects, or wherein the
10 format converter (120) is configured to directly derive DirAC parameters, and wherein the
format combiner (140) is configured to combine the DirAC parameters to obtain the
combined audio scene.
A further embodiment according to the invention provides apparatus, wherein the format
15 converter (120) comprises, a DirAC analyzer (180) for a first order Ambisonics or a high
order Ambisonics input format or a multi-channel signal format, a metadata converter (150,
125, 126, 148) for converting object metadata into DirAC metadata or for converting a multichannel
signal having a time-invariant position into the DirAC metadata, and a metadata
combiner (144) for combining individual DirAC metadata streams or combining direction of
20 arrival metadata from several streams by a weighted addition, the weighting of the weighted
addition being done in accordance to energies of associated pressure signal energies, or
for combining diffuseness metadata from several streams by a weighted addition, the
weighting of the weighted addition being done in accordance with energies of associated
pressure signal energies, or wherein the metadata combiner (144) is configured to
25 calculate, for a time/frequency bin of the first description of the first scene, an energy value,
and direction of arrival value, and to calculate, for the time/frequency bin of the second
description of the second scene, an energy value and a direction of arrival value, and
wherein the format combiner (140) is configured to multiply the first energy to the first
direction of arrival value and to add a multiplication result of the second energy value and
30 the second direction of arrival value to obtain the combined direction of arrival value or,
alternatively, to select the direction of arrival value among the first direction of arrival value
and the second direction of arrival value that is associated with the higher energy as the
combined direction of arrival value.
35 A further embodiment according to the invention provides apparatus, further comprising an
output interface (200, 300) for adding to the combined format, a separate object description
for an audio object, the object description comprising at least one of a direction, a distance,
a diffuseness or any other object attribute, wherein the object has a single direction
48
throughout all frequency bands and is either static or moving slower than a velocity
threshold.
A further embodiment according to the invention provides method for generating a
description of a combined audio scene, comprising, receiving a first description 5 of a first
scene in a first format and receiving a second description of a second scene in a second
format, wherein the second format is different from the first format, converting the first
description into a common format and converting the second description into the common
format, when the second format is different from the common format, and combining the
10 first description in the common format and the second description in the common format to
obtain the combined audio scene.
A further embodiment according to the invention provides computer program for performing,
when running on a computer or a processor, the methods discussed herein above.
15
A further embodiment according to the invention provides apparatus for performing a
sythesis of a plurality of audio scenes, comprising, an input interface (100) for receiving a
first DirAC description of a first scene and for receiving a second DirAC description of a
second scene and one or more transport channels, and a DirAC synthesizer (220) for
20 synthesizing the plurality of audio scenes in a spectral domain to obtain a spectral domain
audio signal representing the plurality of audio scenes, and a spectrum-time converter (240)
for converting the spectral domain audio signal into a time-domain.
Another embodiment according to the invention provides apparatus, wherein the DirAC
25 synthesizer comprises, a scene combiner (221) for combining the first DirAC description
and the second DirAC description into a combined DirAC description, and a DirAC renderer
(222) for rendering the combined DirAC description using one or more transport channels
to obtain the spectral domain audio signal, or wherein the scene combiner (221) is
configured to calculate, for a time/frequency bin of the first description of the first scene, an
30 energy value, and direction of arrival value, and to calculate, for the time/frequency bin of
the second description of the second scene, an energy value and a direction of arrival value,
and wherein the scene combiner (221) is configured to multiply the first energy to the first
direction of arrival value and to add a multiplication result of the second energy value and
the second direction of arrival value to obtain the combined direction of arrival value or,
35 alternatively, to select the direction of arrival value among the first direction of arrival value
and the second direction of arrival value that is associated with the higher energy as the
combined direction of arrival value.
49
Another embodiment according to the invention provides apparatus, wherein the input
interface (100) is configured to receive, for a DirAC description, a separate transport
channel and separate DirAC metadata, wherein the DirAC synthesizer (220) is configured
to render each description using the transport channel and the metadata for the
corresponding DirAC description to obtain a spectral domain audio 5 signal for each
description, and to combine the spectral domain audio signal for each description to obtain
the spectral domain audio signal.
Another embodiment according to the invention provides apparatus, wherein the input
10 interface (100) is configured to receive extra audio object metadata for an audio object, and
wherein the DirAC synthesizer (220) is configured to selectively manipulate the extra audio
object metadata or object data related to the metadata to perform a directional filtering
based on object data included in the object metadata or based on user-given direction
15 information, or wherein the DirAC synthesizer (220) is configured for performing, in the
spectral domain a zero-phase gain function (226), the zero-phase gain function depending
upon a direction of an audio object, wherein the direction is contained in a bitstream if
directions of objects are transmitted as side information, or wherein the direction is received
from a user interface.
20
Another embodiment according to the invention provides method for performing a synthesis
of a plurality of audio scenes, comprising, receiving a first DirAC description of a first scene
and receiving a second DirAC description of a second scene and one or more transport
channels, and synthesizing the plurality of audio scenes in a spectral domain to obtain a
25 spectral domain audio signal representing the plurality of audio scenes, and spectral-time
converting the spectral domain audio signal into a time-domain.
Another embodiment according to the invention provides computer program for performing,
when running on a computer or a processor, the methods discussed herein above.
30
Another embodiment according to the invention provides audio data converter, comprising,
an input interface (100) for receiving an object description of an audio object having audio
object metadata, a metadata converter (150, 125, 126, 148) for converting the audio object
metadata into DirAC metadata, and an output interface (300) for transmitting or storing the
35 DirAC metadata.
50
Another embodiment according to the invention provides audio data converter, in which the
audio object metadata has an object position, and wherein the DirAC metadata has a
direction of arrival with respect to a reference position.
Another embodiment according to the invention provides audio data converter, 5 wherein the
metadata converter (150, 125, 126, 148) is configured to convert DirAC parameters derived
from the object data format into pressure/velocity data and wherein the metadata converter
(150, 125, 126, 148) is configured to apply a DirAC analysis to the pressure/velocity data.
10 Another embodiment according to the invention provides audio data converter, wherein the
input interface (100) is configured to receive a plurality of audio object descriptions, wherein
the metadata converter (150, 125, 126, 148) is configured to convert each object metadata
description into an individual DirAC data description, and wherein the metadata converter
(150, 125, 126, 148) is configured to combine the individual DirAC metadata descriptions
15 to obtain a combined DirAC description as the DirAC metadata.
Another embodiment according to the invention provides audio data converter, wherein the
metadata converter (150, 125, 126, 148) is configured to combine the individual DirAC
metadata descriptions, each metadata description comprising direction of arrival metadata
20 or direction of arrival metadata and diffuseness metadata, by individually combining the
direction of arrival metadata from different metadata descriptions by a weighted addition,
wherein the weighting of the weighted addition is being done in accordance with energies
of associated pressure signal energies, or by combining diffuseness metadata from the
different DirAC metadata descriptions by a weighted addition, the weighting of the weighted
25 addition being done in accordance with energies of associated pressure signal energies,
or, alternatively, to select the direction of arrival value among the first direction of arrival
value and the second direction of arrival value that is associated with the higher energy as
the combined direction of arrival value.
30 Another embodiment according to the invention provides audio data converter, wherein the
input interface (100) is configured to receive, for each audio object, an audio object wave
form signal in addition to this object metadata, wherein the audio data converter further
comprises a downmixer (163) for downmixing the audio object wave form signals into one
or more transport channels, and wherein the output interface (300) is configured to transmit
35 or store the one or more transport channels in association with the DirAC metadata.
Another embodiment according to the invention provides method for performing an audio
data conversion, comprising, receiving an object description of an audio object having audio
51
object metadata, converting the audio object metadata into DirAC metadata, and
transmitting or storing the DirAC metadata.
Another embodiment according to the invention provides computer program for performing,
when running on a computer or a processor, the methods discussed 5 herein above.
Another embodiment according to the invention provides audio scene encoder, comprising,
an input interface (100) for receiving a DirAC description of an audio scene having DirAC
metadata and for receiving an object signal having object metadata, a metadata generator
10 (400) for generating a combined metadata description comprising the DirAC metadata and
the object metadata, wherein the DirAC metadata comprises a direction of arrival for
individual time-frequency tiles and the object metadata comprises a direction or additionally
a distance or a diffuseness of an individual object.
15 Another embodiment according to the invention provides audio scene encoder, wherein the
input interface (100) is configured for receiving a transport signal associated with the DirAC
description of the audio scene and wherein the input interface (100) is configured for
receiving an object wave form signal associated with the object signal, and wherein the
audio scene encoder further comprises a transport signal encoder (170) for encoding the
20 transport signal and the object wave form signal.
Another embodiment according to the invention provides audio scene encoder, wherein the
metadata generator (400) comprises a metadata converter (150, 125, 126, 148).
25 Another embodiment according to the invention provides an audio scene encoder, wherein
the metadata generator (400) is configured to generate, for the object metadata, a single
broadband direction per time and wherein the metadata generator is configured to refresh
the single broadband direction per time less frequently than the DirAC metadata.
30 Another embodiment according to the invention provides method of encoding an audio
scene, comprising, receiving a DirAC description of an audio scene having DirAC metadata
and receiving an object signal having audio object metadata, and generating a combined
metadata description comprising the DirAC metadata and the object metadata, wherein the
DirAC metadata comprises a direction of arrival for individual time-frequency tiles and
35 wherein the object metadata comprises a direction or, additionally, a distance or a
diffuseness of an individual object.
52
Another embodiment according to the invention provides computer program for performing,
when running on a computer or a processor, the methods discussed herein above.
Another embodiment according to the invention provides apparatus for performing a
synthesis of audio data, comprising, an input interface (100) for 5 receiving a DirAC
description of one or more audio objects or a multi-channel signal or a first order Ambisonics
signal or a high order Ambisonics signal, wherein the DirAC description comprises position
information of the one or more objects or side information for the first order Ambisonics
signal or the high order Ambisonics signal or a position information for the multi-channel
10 signal as side information or from a user interface, a manipulator (500) for manipulating the
DirAC description of the one or more audio objects, the multi-channel signal, the first order
Ambisonics signal or the high order Ambisonics signal to obtain a manipulated DirAC
description, and a DirAC synthesizer (220, 240) for synthesizing the manipulated DirAC
description to obtain synthesized audio data.
15
Another embodiment according to the invention provides apparatus, wherein the DirAC
synthesizer (220, 240) comprises a DirAC renderer (222) for performing a DirAC rendering
using the manipulated DirAC description to obtain a spectral domain audio signal, and a
spectral-time converter (240) to convert the spectral domain audio signal into a time20
domain.
Another embodiment according to the invention provides apparatus, wherein the
manipulator (500) is configured to perform a position-dependent weighting operation prior
to DirAC rendering.
25
Another embodiment according to the invention provides apparatus, wherein the DirAC
synthesizer (220, 240) is configured to output a plurality of objects or a first order
Ambisonics signal or a high order Ambisonics signal or a multi-channel signal, and wherein
the DirAC synthesizer (220, 240) is configured to use a separate spectral-time converter
30 (240) for each object or each component of the first order Ambisonics signal or the high
order Ambisonics signal or for each channel of the multi-channel signal.
Another embodiment according to the invention provides method for performing a synthesis
of audio data, comprising, receiving a DirAC description of one or more audio objects or a
35 multi-channel signal or a first order Ambisonics signal or a high order Ambisonics signal,
wherein the DirAC description comprising position information of the one or more objects
or of the multi-channel signal or additional information for the first order Ambisonics signal
or the high order Ambisonics signal as side information or for a user interface, manipulating
53
the DirAC description to obtain a manipulated DirAC description, and synthesizing the
manipulated DirAC description to obtain synthesized audio data.
Another embodiment according to the invention provides computer program for performing,
when running on a computer or a processor, the methods discussed 5 herein above.
The above described embodiments are merely illustrative for the principles of the present
invention. It is understood that modifications and variations of the arrangements and the
details described herein will be apparent to others skilled in the art. It is the intent, therefore,
10 to be limited only by the scope of the impending patent claims and not by the specific details
presented by way of description and explanation of the embodiments herein.
54
We Claim:
1. Audio data converter, comprising:
an input interface (100) for receiving an object description of an audio 5 object having
audio object metadata;
a metadata converter (150, 125, 126, 148) for converting the audio object metadata
into DirAC metadata; and
10
an output interface (300) for transmitting or storing the DirAC metadata.
2. Audio data converter of claim 1, in which the audio object metadata has an object
position, and wherein the DirAC metadata has a direction of arrival with respect to a
15 reference position.
3. Audio data converter of one of claims 1 or 2,
wherein the metadata converter (150, 125, 126, 148) is configured to convert DirAC
20 parameters derived from the object data format into pressure/velocity data, and
wherein the metadata converter (150, 125, 126, 148) is configured to apply a DirAC
analysis to the pressure/velocity data.
25 4. Audio data converter in accordance with one of claims 1 to 3,
wherein the input interface (100) is configured to receive a plurality of audio object
descriptions,
30 wherein the metadata converter (150, 125, 126, 148) is configured to convert each
object metadata description into an individual DirAC data description, and
wherein the metadata converter (150, 125, 126, 148) is configured to combine the
individual DirAC metadata descriptions to obtain a combined DirAC description as
35 the DirAC metadata.
5. Audio data converter in accordance with claim 4, wherein the metadata converter
(150, 125, 126, 148) is configured to combine the individual DirAC metadata
55
descriptions, each metadata description comprising direction of arrival metadata, by
individually combining the direction of arrival metadata from different metadata
descriptions by a weighted addition, wherein the weighting of the weighted addition
is being done in accordance with energies of associated pressure signal energies.
5
6. Audio data converter in accordance with claim 4, wherein the metadata converter
(150, 125, 126, 148) is configured to combine the individual DirAC metadata
descriptions, each metadata description comprising direction of arrival metadata and
diffuseness metadata, by individually combining the direction of arrival metadata
10 from different metadata descriptions by a weighted addition, wherein the weighting
of the weighted addition is being done in accordance with energies of associated
pressure signal energies, and by combining the diffuseness metadata from the
different DirAC metadata descriptions by a weighted addition, the weighting of the
weighted addition being done in accordance with energies of associated pressure
15 signal energies.
7. Audio data converter in accordance with claim 4, wherein the metadata converter
(150, 125, 126, 148) is configured to combine the individual DirAC metadata
descriptions, each metadata description comprising direction of arrival metadata or
20 direction of arrival metadata and diffuseness metadata, by selecting the direction of
arrival value among a first direction of arrival value of a first DirAC metadata
description and a second direction of arrival value of a second DirAC metadata
description that is associated with a higher energy of an associated pressure signal
energy as a combined direction of arrival value.
25
8. Audio data converter is accordance with one of claims 1 to 7,
wherein the input interface (100) is configured to receive, for each audio object, an
audio object wave form signal in addition to this object metadata,
30
wherein the audio data converter further comprises a downmixer (163) for
downmixing the audio object wave form signals into one or more transport channels,
and
35 wherein the output interface (300) is configured to transmit or store the one or more
transport channels in association with the DirAC metadata.
9. Method for performing an audio data conversion, comprising:
56
receiving an object description of an audio object having audio object metadata;
converting the audio object metadata into DirAC metadata; and
5
transmitting or storing the DirAC metadata.
10. Computer program for performing, when running on a computer or a processor, the
method of claim 9.
10
11. Audio scene encoder, comprising:
an input interface (100) for receiving a DirAC description of an audio scene having
DirAC metadata and for receiving an object signal having object metadata;
15
a metadata generator (400) for generating a combined metadata description
comprising the DirAC metadata and the object metadata, wherein the DirAC
metadata comprises a direction of arrival for individual time-frequency tiles and the
object metadata comprises a direction or additionally a distance or a diffuseness of
20 an individual object, wherein the metadata generator (400) comprises a metadata
converter (150, 125, 126, 148) as described in any of the claims 1 to 7.
12. Audio scene encoder of claim 11, wherein the input interface (100) is configured for
receiving a transport signal associated with the DirAC description of the audio scene
25 and wherein the input interface (100) is configured for receiving an object wave form
signal associated with the object signal, and
wherein the audio scene encoder further comprises a transport signal encoder (170)
for encoding the transport signal and the object wave form signal.
30
13. An audio scene encoder of one of claims 11 to 12,
wherein the metadata generator (400) is configured to generate, for the object
metadata, a single broadband direction per time and wherein the metadata
35 generator is configured to refresh the single broadband direction per time less
frequently than the DirAC metadata.
14. Method of encoding an audio scene, comprising:
57
receiving a DirAC description of an audio scene having DirAC metadata and
receiving an object signal having audio object metadata; and
generating a combined metadata description comprising the 5 DirAC metadata and
the object metadata, wherein the DirAC metadata comprises a direction of arrival for
individual time-frequency tiles and wherein the object metadata comprises a
direction or, additionally, a distance or a diffuseness of an individual object, wherein
the generating comprises using a metadata generator (400) comprising a metadata
10 converter (150, 125, 126, 148) as described in any of the claims 1 to 7.
15. Computer program for performing, when running on a computer or a processor, the
method of claim 14.
15 16. Apparatus for performing a synthesis of audio data, comprising:
an input interface (100) for receiving a DirAC description of one or more audio
objects or a multi-channel signal or a first order Ambisonics signal or a high order
Ambisonics signal, wherein the DirAC description comprises position information of
20 the one or more objects or side information for the first order Ambisonics signal or
the high order Ambisonics signal or a position information for the multi-channel
signal as side information or from a user interface;
a manipulator (500) for manipulating the DirAC description of the one or more audio
25 objects, the multi-channel signal, the first order Ambisonics signal or the high order
Ambisonics signal to obtain a manipulated DirAC description; and
a DirAC synthesizer (220, 240) for synthesizing the manipulated DirAC description
to obtain synthesized audio data.
30
17. Apparatus of claim 16,
wherein the DirAC synthesizer (220, 240) comprises a DirAC renderer (222) for
performing a DirAC rendering using the manipulated DirAC description to obtain a
35 spectral domain audio signal; and
a spectral-time converter (240) to convert the spectral domain audio signal into a
time-domain.
58
18. Apparatus of claim 16 or 17,
wherein the manipulator (500) is configured to perform a position-dependent
weighting operation prior to 5 DirAC rendering.
19. Apparatus of one of claims 16 to 18,
wherein the DirAC synthesizer (220, 240) is configured to output a plurality of objects
10 or a first order Ambisonics signal or a high order Ambisonics signal or a multichannel
signal, and wherein the DirAC synthesizer (220, 240) is configured to use a
separate spectral-time converter (240) for each object or each component of the first
order Ambisonics signal or the high order Ambisonics signal or for each channel of
the multi-channel signal.
15
20. Method for performing a synthesis of audio data, comprising:
receiving a DirAC description of one or more audio objects or a multi-channel signal
or a first order Ambisonics signal or a high order Ambisonics signal, wherein the
20 DirAC description comprising position information of the one or more objects or of
the multi-channel signal or additional information for the first order Ambisonics signal
or the high order Ambisonics signal as side information or for a user interface;
manipulating the DirAC description to obtain a manipulated DirAC description; and
synthesizing the manipulated DirAC description to obtain synthesized audio data.
21. Computer program for performing, when running on a computer or a processor, the
method of claim 20.

Documents

Application Documents

# Name Date
1 202128059438-STATEMENT OF UNDERTAKING (FORM 3) [20-12-2021(online)].pdf 2021-12-20
2 202128059438-REQUEST FOR EXAMINATION (FORM-18) [20-12-2021(online)].pdf 2021-12-20
3 202128059438-FORM 18 [20-12-2021(online)].pdf 2021-12-20
4 202128059438-FORM 1 [20-12-2021(online)].pdf 2021-12-20
5 202128059438-FIGURE OF ABSTRACT [20-12-2021(online)].jpg 2021-12-20
6 202128059438-DRAWINGS [20-12-2021(online)].pdf 2021-12-20
7 202128059438-DECLARATION OF INVENTORSHIP (FORM 5) [20-12-2021(online)].pdf 2021-12-20
8 202128059438-COMPLETE SPECIFICATION [20-12-2021(online)].pdf 2021-12-20
9 Abstract1.jpg 2021-12-23
10 202128059438.pdf 2021-12-23
11 202128059438-Proof of Right [05-01-2022(online)].pdf 2022-01-05
12 202128059438-FORM-26 [11-02-2022(online)].pdf 2022-02-11
13 202128059438-Information under section 8(2) [28-03-2022(online)].pdf 2022-03-28
14 202128059438-FER.pdf 2022-04-04
15 202128059438-Information under section 8(2) [19-05-2022(online)].pdf 2022-05-19
16 202128059438-FORM 3 [19-05-2022(online)].pdf 2022-05-19
17 202128059438-certified copy of translation [20-06-2022(online)].pdf 2022-06-20
18 202128059438-Certified Copy of Priority Document [20-06-2022(online)].pdf 2022-06-20
19 202128059438-Information under section 8(2) [15-07-2022(online)].pdf 2022-07-15
20 202128059438-Information under section 8(2) [30-08-2022(online)].pdf 2022-08-30
21 202128059438-FORM 3 [30-08-2022(online)].pdf 2022-08-30
22 202128059438-FORM 4(ii) [23-09-2022(online)].pdf 2022-09-23
23 202128059438-Information under section 8(2) [08-12-2022(online)].pdf 2022-12-08
24 202128059438-OTHERS [03-01-2023(online)].pdf 2023-01-03
25 202128059438-FER_SER_REPLY [03-01-2023(online)].pdf 2023-01-03
26 202128059438-DRAWING [03-01-2023(online)].pdf 2023-01-03
27 202128059438-CLAIMS [03-01-2023(online)].pdf 2023-01-03
28 202128059438-Information under section 8(2) [14-02-2023(online)].pdf 2023-02-14
29 202128059438-FORM 3 [14-02-2023(online)].pdf 2023-02-14
30 202128059438-Information under section 8(2) [10-07-2023(online)].pdf 2023-07-10
31 202128059438-Information under section 8(2) [09-08-2023(online)].pdf 2023-08-09
32 202128059438-FORM 3 [09-08-2023(online)].pdf 2023-08-09
33 202128059438-Information under section 8(2) [13-10-2023(online)].pdf 2023-10-13
34 202128059438-Information under section 8(2) [09-02-2024(online)].pdf 2024-02-09
35 202128059438-FORM 3 [09-02-2024(online)].pdf 2024-02-09
36 202128059438-US(14)-HearingNotice-(HearingDate-06-05-2024).pdf 2024-03-27
37 202128059438-Information under section 8(2) [02-04-2024(online)].pdf 2024-04-02
38 202128059438-Correspondence to notify the Controller [01-05-2024(online)].pdf 2024-05-01
39 202128059438-FORM-26 [03-05-2024(online)].pdf 2024-05-03
40 202128059438-Written submissions and relevant documents [20-05-2024(online)].pdf 2024-05-20
41 202128059438-PatentCertificate24-06-2024.pdf 2024-06-24
42 202128059438-IntimationOfGrant24-06-2024.pdf 2024-06-24

Search Strategy

1 searchE_31-03-2022.pdf

ERegister / Renewals

3rd: 02 Aug 2024

From 01/10/2020 - To 01/10/2021

4th: 02 Aug 2024

From 01/10/2021 - To 01/10/2022

5th: 02 Aug 2024

From 01/10/2022 - To 01/10/2023

6th: 02 Aug 2024

From 01/10/2023 - To 01/10/2024

7th: 02 Aug 2024

From 01/10/2024 - To 01/10/2025

8th: 30 Sep 2025

From 01/10/2025 - To 01/10/2026