Abstract: An apparatus for providing an upmix signal representation on the basis of a downmix signal representation and an object-related parametric information, which are included in a bitstream representation of an audio content, in independence on a user-specified rendering matrix, the apparatus comprises a distortion limiter configured to obtain a modified rendering matrix using a linear combination of a user-specified rendering matrix in a target rendering matrix in dependence on a linear combination parameter. The apparatus also comprises a signal processor configured to obtain the upmix signal representation on the basis of the downmix signal representation and the object-related parametric information using the modified rendering matrix. The apparatus is also configured to evaluate a bitstream element representing the linear combination parameter in order to obtain the linear combination parameter.
Apparatus for providing an upmix signal representation on the basis of the
downmix signal representation, apparatus for providing a bitstream representing
a multi-channel audio signal, methods, computer programs and bitstream
representing a multi-channel audio signal using a linear combination parameter
Technical Field
Embodiments according to the invention are related to an apparatus for providing an
upmix signal representation on the basis of a downmix signal representation and an
object-related parametric information, which are included in a bitstream representation
of an audio content, and in dependence on a user-specified rendering matrix.
Other embodiments according to the invention are related to an apparatus for providing
a bitstream representing a multi-channel audio signal.
Other embodiments according to the invention are related to a method for providing an
upmix signal representation on the basis of a downmix signal representation and an
object-related parametric information which are included in a bitstream representation
of the audio content, and in dependence on a user-specified rendering matrix.
Other embodiments according to the invention are related to a method for providing a
bitstream representing a multi-channel audio signal.
Other embodiments according to the invention are related to a computer program
performing one of said methods.
Another embodiment according to the invention is related to a bitstream representing a
multi-channel audio signal.
Background of the Invention
In the art of audio processing, audio transmission and audio storage there is an
increasing desire to handle multi-channel contents in order to improve the hearing
impression. Usage of a multi-channel audio content brings along significant
improvements for the user. For example, a 3-dimensional hearing impression can be
obtained, which brings along an improved user satisfaction in entertainment
applications. However, multi-channel audio contents are also useful in professional
environments, for example, telephone conferencing applications, because the speaker
intelligibility can be improved by using a multi-channel audio playback.
However, it is also desirable to have a good trade-off between audio quality and bitrate
requirements in order to avoid excessive resource consumption in low-cost or
professional multi-channel applications.
Parametric techniques for the bitrate-efficient transmission and/or storage of audio
scenes containing multiple audio objects have recently been proposed. For example, a
binaural cue coding, which is described, for example, in reference [1], and a parametric
joint-coding of audio sources, which is described, for example, in reference [2], have
been proposed. Also, an MPEG spatial audio object coding (SAOC) has been proposed,
which is described, for example, in references [3] and [4]. MPEG spatial audio object
coding is currently under standardization, and described in non-pre-published reference
[5].
These techniques aim at perceptually reconstructing the desired output scene rather than
by a wave form match.
However, in combination with user interactivity at the receiving side, such techniques
may lead to a low audio quality of the output audio signals if extreme object rendering
is performed. This is described, for example, in reference [6].
In the following, such systems will be described, and it should be noted that the basic
concepts also apply to the embodiments of the invention.
Fig. 8 shows a system overview of such a system (here: MPEG SAOC). The MPEG
SAOC system 800 shown in Fig. 8 comprises an SAOC encoder 810 and an SAOC
decoder 820. The SAOC encoder 810 receives a plurality of object signals X1 to xn,
which may be represented, for example, as time-domain signals or as time-frequency-
domain signals (for example, in the form of a set of transform coefficients of a Fourier-
type transform, or in the form of QMF subband signals). The SAOC encoder 810
typically also receives downmix coefficients d1 to dN, which are associated with the
object signals x1 to xn. Separate sets of downmix coefficients may be available for each
channel of the downmix signal. The SAOC encoder 810 is typically configured to
obtain a channel of the downmix signal by combining the object signals x1 to xn in
accordance with the associated downmix coefficients d1 to dN. Typically, there are less
downmix channels than object signals x1 to xN. In order to allow (at least
approximately) for a separation (or separate treatment) of the object signals at the side
of the SAOC decoder 820, the SAOC encoder 810 provides both the one or more
downmix signals (designated as downmix channels) 812 and a side information 814.
The side information 814 describes characteristics of the object signals x1 to xN, in order
to allow for a decoder-sided object-specific processing.
The SAOC decoder 820 is configured to receive both the one or more downmix signals
812 and the side information 814. Also, the SAOC decoder 820 is typically configured
to receive a user interaction information and/or a user control information 822, which
describes a desired rendering setup. For example, the user interaction information/user
control information 822 may describe a speaker setup and the desired spatial placement
of the objects which provide the object signals x1 to xN
The SAOC decoder 820 is configured to provide, for example, a plurality of decoded
upmix channel signals y1 to yM The upmix channel signals may for example be
associated with individual speakers of a multi-speaker rendering arrangement. The
SAOC decoder 820 may, for example, comprise an object separator 820a, which is
configured to reconstruct, at least approximately, the object signals x1 to xN on the basis
of the one or more downmix signals 812 and the side information 814, thereby
obtaining reconstructed object signals 820b. However, the reconstructed object signals
820b may deviate somewhat from the original object signals x1 to xN, for example,
because the side information 814 is not quite sufficient for a perfect reconstruction due
to the bitrate constraints. The SAOC decoder 820 may further comprise a mixer 820c,
which may be configured to receive the reconstructed object signals 820b and the user
interaction information/user control information 822, and to provide, on the basis
thereof, the upmix channel signals y1 to yM. The mixer 820 may be configured to use the
user interaction information /user control information 822 to determine the contribution
of the individual reconstructed object signals 820b to the upmix channel signals y1 to
yM. The user interaction information/user control information 822 may, for example,
comprise rendering parameters (also designated as rendering coefficients), which
determine the contribution of the individual reconstructed object signals 822 to the
upmix channel signals y1 to yM.
However, it should be noted that in many embodiments, the object separation, which is
indicated by the object separator 820a in Fig. 8, and the mixing, which is indicated by
the mixer 820c in Fig. 8, are performed in single step. For this purpose, overall
parameters may be computed which describe a direct mapping of the one or more
downmix signals 812 onto the upmix channel signals y1 to yM These parameters may be
computed on the basis of the side information and the user interaction information/user
control information 820.
Taking reference now to Figs. 9a, 9b and 9c, different apparatus for obtaining an upmix
signal representation on the basis of a downmix signal representation and object-related
side information will be described. Fig. 9a shows a block schematic diagram of a
MPEG SAOC system 900 comprising an SAOC decoder 920. The SAOC decoder 920
comprises, as separate functional blocks, an object decoder 922 and a mixer/renderer
926. The object decoder 922 provides a plurality of reconstructed object signals 924 in
dependence on the downmix signal representation (for example, in the form of one or
more downmix signals represented in the time domain or in the time-frequency-domain)
and object-related side information (for example, in the form of object meta data). The
mixer/renderer 924 receives the reconstructed object signals 924 associated with a
plurality of N objects and provides, on the basis thereof, one or more upmix channel
signals 928. In the SAOC decoder 920, the extraction of the object signals 924 is
performed separately from the mixing/rendering which allows for a separation of the
object decoding functionality from the mixing/rendering functionality but brings along
a relatively high computational complexity.
Taking reference now to Fig. 9b, another MPEG SAOC system 930 will be briefly
discussed, which comprises an SAOC decoder 950, The SAOC decoder 950 provides a
plurality of upmix channel signals 958 in dependence on a downmix signal
representation (for example, in the form of one or more downmix signals) and an
object-related side information (for example, in the form of object meta data). The
SAOC decoder 950 comprises a combined object decoder and mixer/renderer, which is
configured to obtain the upmix channel signals 958 in a joint mixing process without a
separation of the object decoding and the mixing/rendering, wherein the parameters for
said joint upmix process are dependent both on the object-related side information and
the rendering information. The joint upmix process depends also on the downmix
information, which is considered to be part of the object-related side information.
To summarize the above, the provision of the upmix channel signals 928, 958 can be
performed in a one step process or a two step process.
Taking reference now to Fig. 9c, an MPEG SAOC system 960 will be described. The
SAOC system 960 comprises an SAOC to MPEG Surround transcoder 980, rather than
an SAOC decoder.
The SAOC to MPEG Surround transcoder comprises a side information transcoder 982,
which is configured to receive the object-related side information (for example, in the
form of object meta data) and, optionally, information on the one or more downmix
signals and the rendering information. The side information transcoder is also
configured to provide an MPEG Surround side information (for example, in the form of
an MPEG Surround bitstream) on the basis of a received data. Accordingly, the side
information transcoder 982 is configured to transform an object-related (parametric)
side information, which is relieved from the object encoder, into a channel-related
(parametric) side information, taking into consideration the rendering information and,
optionally, the information about the content of the one or more downmix signals.
Optionally, the SAOC to MPEG Surround transcoder 980 may be configured to
manipulate the one or more downmix signals, described, for example, by the downmix
signal representation, to obtain a manipulated downmix signal representation 988.
However, the downmix signal manipulator 986 may be omitted, such that the output
downmix signal representation 988 of the SAOC to MPEG Surround transcoder 980 is
identical to the input downmix signal representation of the SAOC to MPEG Surround
transcoder. The downmix signal manipulator 986 may, for example, be used if the
channel-related MPEG Surround side information 984 would not allow to provide a
desired hearing impression on the basis of the input downmix signal representation of
the SAOC to MPEG Surround transcoder 980, which may be the case in some rendering
constellations.
Accordingly, the SAOC to MPEG Surround transcoder 980 provides the downmix
signal representation 988 and the MPEG Surround bitstream 984 such that a plurality of
upmix channel signals, which represent the audio objects in accordance with the
rendering information input to the SAOC to MPEG Surround transcoder 980 can be
generated using an MPEG Surround decoder which receives the MPEG Surround
bitstream 984 and the downmix signal representation 988.
To summarize the above, different concepts for decoding SAOC-encoded audio signals
can be used. In some cases, a SAOC decoder is used, which provides upmix channel
signals (for example, upmix channel signals 928, 958) in dependence on the downmix
signal representation and the object-related parametric side information. Examples for
this concept can be seen in Figs. 9a and 9b. Alternatively, the SAOC-encoded audio
information may be transcoded to obtain a downmix signal representation (for example,
a downmix signal representation 988) and a channel-related side information (for
example, the channel-related MPEG Surround bitstream 984), which can be used by an
MPEG Surround decoder to provide the desired upmix channel signals.
In the MPEG SAOC system 800, a system overview of which is given in Fig. 8, the
general processing is carried out in a frequency selective way and can be described as
follows within each frequency band:
• N input audio object signals x1 to xn are downmixed as part of the SAOC encoder
processing. For a mono downmix, the downmix coefficients are denoted by d1 to dN.
In addition, the SAOC encoder 810 extracts side information 814 describing the
characteristics of the input audio objects. For MPEG SAOC, the relations of the
object powers with respect to each other are the most basic form of such a side
information.
• Downmix signal (or signals) 812 and side information 814 are transmitted and/or
stored. To this end, the downmix audio signal may be compressed using well-known
perceptual audio coders such as MPEG-1 Layer II or III (also known as ".mp3"),
MPEG Advanced Audio Coding (AAC), or any other audio coder.
• On the receiving end, the SAOC decoder 820 conceptually tries to restore the
original object signal ("object separation") using the transmitted side information
814 (and, naturally, the one or more downmix signals 812). These approximated
object signals (also designated as reconstructed object signals 820b) are then mixed
into a target scene represented by M audio output channels (which may, for
example, be represented by the upmix channel signals y1 to ym) using a rendering
matrix. For a mono output, the rendering matrix coefficients are given by r1 to rN
• Effectively, the separation of the object signals is rarely executed (or even never
executed), since both the separation step (indicated by the object separator 820a) and
the mixing step (indicated by the mixer 820c) are combined into a single transcoding
step, which often results in an enormous reduction in computational complexity.
It has been found that such a scheme is tremendously efficient, both in terms of
transmission bitrate (it is only necessary to transmit a few downmix channels plus some
side information instead of N discrete object audio signals or a discrete system) and
computational complexity (the processing complexity relates mainly to the number of
output channels rather than the number of audio objects). Further advantages for the
user on the receiving end include the freedom of choosing a rendering setup of his/her
choice (mono, stereo, surround, virtualized headphone playback, and so on) and the
feature of user interactivity: the rendering matrix, and thus the output scene, can be set
and changed interactively by the user according to will, personal preference or other
criteria. For example, it is possible to locate the talkers from one group together in one
spatial area to maximize discrimination from other remaining talkers. This interactivity
is achieved by providing a decoder user interface:
For each transmitted sound object, its relative level and (for non-mono rendering)
spatial position of rendering can be adjusted. This may happen in real-time as the user
changes the position of the associated graphical user interface (GUI) sliders (for
example: object level = +5dB, object position = -30deg).
However, it has been found that the decoder-sided choice of parameters for the
provision of the upmix signal representation (e.g. the upmix channel signals y1 to yM)
brings along audible degradations in some cases.
In view of this situation, it is the objective of the present invention to create a concept
which allows for reducing or even avoiding audible distortion when providing an upmix
signal representation (for example, in the form of upmix channel signals y1 to yM)-
Summarv of the Invention
An embodiment according to the invention creates an apparatus for providing an upmix
signal representation on the basis of a downmix signal representation and an object-
related parametric information, which are included in a bitstream representation of an
audio content, and in dependence on a user-specified rendering matrix. The apparatus
comprises a distortion limiter configured to obtain a modified rendering matrix using a
linear combination of a user-specified rendering matrix and a target rendering matrix in
dependence on a linear combination parameter. The apparatus also comprises a signal
processor configured to obtain the upmix signal representation on the basis of the
downmix signal representation and the object-related parametric information using the
modified rendering matrix. The apparatus is configured to evaluate a bitstream element
representing the linear combination parameter in order to obtain the linear combination
parameter.
This embodiment according to the invention is based on the key idea that audible
distortions of the upmix signal representation can be reduced or even avoided with low
computational complexity by performing a linear combination of a user-specified
rendering matrix and the target rendering matrix in dependence on a linear combination
parameter, which is extracted from the bitstream representation of the audio content,
because a linear combination can be performed efficiently, and because the execution of
the demanding task of determining the linear combination parameter can be performed
at the side of the audio signal encoder where there is typically more computational
power available than at the side of the audio signal decoder (apparatus for providing an
upmix signal representation).
Accordingly, the above-discussed concept allows to obtain a modified rendering matrix,
which results in reduced audible distortions even for an inappropriate choice of the user-
specified rendering matrix, without adding any significant complexity to the apparatus
for providing an upmix signal representation. In particular, it may even be unnecessary
to modify the signal processor when compared to an apparatus without a distortion
limiter, because the modified rendering matrix constitutes an input quantity to the signal
processor and merely replaces the user-specified rendering matrix. In addition, the
inventive concept brings along the advantage that an audio signal encoder can adjust the
distortion limitation scheme, which is applied at the side of the audio signal decoder, in
accordance with requirements specified at the encoder side by simply setting the linear
combination parameter, which is included in the bitstream representation of the audio
content. Accordingly, the audio signal encoder may gradually provide more or less
freedom with respect to the choice of the rendering matrix to the user of the decoder
(apparatus for providing an upmix signal representation) by appropriately choosing the
linear combination parameter. This allows for the adaptation of the audio signal decoder
to the user's expectations for a given service, because for some services a user may
expect a maximum quality (which implies to reduce the user's possibility to arbitrarily
adjust the rendering matrix), while for other services the user may typically expect a
maximum degree of freedom (which implies to increase the impact of the user's
specified rendering matrix onto the result of the linear combination).
To summarize the above, the inventive concept combines high computational efficiency
at the decoder side, which may be particularly important for portable audio decoders,
with the possibility of a simple implementation, without bringing along the need to
modify the signal processor, and also provides a high degree of control to an audio
signal encoder, which may be important to fulfill the user's expectations for different
types of audio services.
In a preferred embodiment, the distortion limiter is configured to obtain the target
rendering matrix such that the target rendering matrix is a distortion-free target
rendering matrix. This brings along the possibility to have a playback scenario in which
there are no distortions or at least hardly any distortions caused by the choice of the
rendering matrix. Also, it has been found that the computation of a distortion-free target
rendering matrix can be performed in a very simple manner in some cases. Further, it
has been found that a rendering matrix, which is chosen in-between a user-specified
rendering matrix and a distortion-free target rendering matrix typically results in a good
hearing impression.
In a preferred embodiment, the distortion limiter is configured to obtain the target
rendering matrix such that the target rendering matrix is a downmix-similar target
rendering matrix. It has been found that the usage of a downmix-similar target rendering
matrix brings along a very low or even minimal degree of distortions. Also, such a
downmix-similar target rendering matrix can be obtained with very low computational
effort, because the downmix-similar target rendering matrix can be obtained by scaling
the entries of the downmix matrix with a common scaling factor and adding some
additional zero entries.
In a preferred embodiment, the distortion limiter is configured to scale an extended
downmix matrix using an energy normalization scalar, to obtain the target rendering
matrix, wherein the extended downmix matrix is an extended version of the downmix
matrix (a row of which downmix matrix describes contributions of a plurality of audio
object signals to the one or more channels of the downmix signal representation),
extended by rows of zero elements, such that a number of rows of the extended
downmix matrix is identical to a rendering constellation described by the user-specified
rendering matrix. Thus, the extended downmix matrix is obtained using a copying of
values from the downmix matrix into the extended downmix matrix, an addition of zero
matrix entries, and a scalar multiplication of all the matrix elements with the same
energy normalization scalar. All of these operations can be performed very efficiently,
such that the target rendering matrix can be obtained fast, even in a very simple audio
decoder.
In a preferred embodiment, the distortion limiter is configured to obtain the target
rendering matrix such that the target rendering matrix is a best-effort target rendering
matrix. Even though this approach is computationally somewhat more demanding than
the usage of a downmix-similar target rendering matrix, the usage of a best-effort target
rendering matrix provides for a better consideration of a user's desired rendering
scenario. Using the best-effort target rendering matrix, a user's definition of the desired
rendering matrix is taken into consideration when determining the target rendering
matrix as far as it is possible without introducing distortions or significant distortions. In
particular, the best-effort target rendering matrix takes into consideration the user's
desired loudness for a plurality of speakers (or channels of the upmix signal
representation). Accordingly, an improved hearing impression may result when using
the best-effort target rendering matrix.
In a preferred embodiment, the distortion limiter is configured to obtain the target
rendering matrix such that the target rendering matrix depends on a downmix matrix
and the user's specified rendering matrix. Accordingly, the target rendering matrix is
relatively close to the user's expectations but still provides for a substantially distortion-
free audio rendering. Thus, the linear combination parameter determines a trade-off
between an approximation of the user's desired rendering and minimization of audible
distortions, wherein the consideration of the user-specified rendering matrix for the
computation of the target rendering matrix provides for a good satisfaction of the user's
desires, even if the linear combination parameter indicates that the target rendering
matrix should dominate the linear combination.
In a preferred embodiment, the distortion limiter is configured to compute a matrix
comprising channel-individual normalization values for a plurality of output audio
channels of the apparatus for providing an upmix signal representation, such that an
energy normalization value for a given output channel of the apparatus describes, at
least approximately, a ratio between a sum of energy rendering values associated with
the given output channel in the user-specified rendering matrix for a plurality of audio
objects, and a sum of energy downmix values for the plurality of audio objects.
Accordingly, a user's expectation with respect to the loudness of the different output
channels of the apparatus can be met to some degree.
In this case the distortion limiter is configured to scale a set of downmix values using an
associated channel-individual energy normalization value, to obtain a set of rendering
values of the target rendering matrix associated with the given output channel.
Accordingly, the relative contribution of a given audio object to an output channel of
the apparatus is identical to the relative contribution of the given audio object to the
downmix signal representation, which allows to substantially avoid audible distortions
which would be caused by a modification of the relative contributions of the audio
objects. Accordingly, each of the output channels of the apparatus is substantially
undistorted. Nevertheless, the user's expectation with respect to a loudness distribution
over a plurality of speakers (or channels of the upmix signal representation) is taken
into consideration, even though details where to place which audio object and/or how to
change relative intensities of the audio objects with respect to each other are left
unconsidered (at least to some degree) in order to avoid distortions which would
possibly be caused by an excessively sharp spatial separation of the audio objects or an
excessive modification of relative intensities of audio objects.
Thus, evaluating the ratio between a sum of energy rendering values (for example,
squares of magnitude rendering values) associated with a given output channel in the
user-specified rendering matrix for a plurality of audio objects and a sum of energy
downmix values for the plurality of audio objects allows to consider all of the output
audio channels, even though the downmix signal representation may comprise of less
channels, while still avoiding distortions which would be caused by a spatial
redistribution of audio objects or by an excessive change of the relative loudness of the
different audio objects.
In a preferred embodiment, the distortion limiter is configured to compute a matrix
describing a channel-individual energy normalization for a plurality of output audio
channels of the apparatus for providing an upmix signal representation in dependence
on the user-specified rendering matrix and a downmix matrix. In this case, the distortion
limiter is configured to apply the matrix describing the channel-individual energy
normalization to obtain a set of rendering coefficients of the target rendering matrix
associated with the given output channel of the apparatus as a linear combination of sets
of downmix values (i.e., values describing a scaling applied to the audio signals of
different audio objects to obtain a channel of the downmix signal) associated with
different channels of the downmix signal representation. Using this concept, a target
rendering matrix, which is well-adapted to the desired user-specified rendering matrix,
can be obtained even if the downmix signal representation comprises more than one
audio channel, while still substantially avoiding distortions. It has been found that the
formation of a linear combination of sets of downmix values results in a set of rendering
coefficients which typically causes only small audible distortions. Nevertheless, it has
been found that it is possible to approximate a user's expectation using such an
approach for deriving the target rendering matrix.
In a preferred embodiment, the apparatus is configured to read an index value
representing the linear combination parameter from the bitstream representation of the
audio content, and to map the index value onto the linear combination parameter using a
parameter quantization table. It has been found that this is a particularly
computationally efficient concept for deriving the linear combination parameter. It has
also been found that this approach brings along a better trade-off between user's
satisfaction and computational complexity when compared to other possible concepts in
which complicated computations, rather than the evaluation of a 1-dimensional mapping
table, are performed.
In a preferred embodiment, the quantization table describes a non-uniform quantization,
wherein smaller values of the linear combination parameter, which describe a stronger
contribution of the user-specified rendering matrix onto the modified rendering matrix,
are quantized with comparatively high resolution and larger values of the linear
combination parameter, which describe a smaller contribution of the user-specified
rendering matrix onto the modified rendering matrix are quantized with comparatively
lower resolution. It has been found that in many cases only extreme settings of the
rendering matrix bring along significant audible distortions. Accordingly, it has been
found that a fine adjustment of the linear combination parameter is more important in
the region of a stronger contribution of the user-specified rendering matrix onto the
target rendering matrix, in order to obtain a setting which allows for an optimal trade-
off between a fulfillment of a user's rendering expectation and a minimization of
audible distortions.
In a preferred embodiment, the apparatus is configured to evaluate a bitstream element
describing a distortion limitation mode. In this case, the distortion limiter is preferably
configured to selectively obtain the target rendering matrix such that the target
rendering matrix is a downmix-similar target rendering matrix or such that the target
rendering matrix is a best-effort target rendering matrix. It has been found that such a
switchable concept provides for an efficient possibility to obtain a good trade-off
between a fulfillment of a user's rendering expectations and a minimization of the
audible distortions for a large number of different audio pieces. This concept also
allows for a good control of an audio signal encoder over the actual rendering at the
decoder side. Consequently, the requirements of a large variety of different audio
services can be fulfilled.
Another embodiment according to the invention creates an apparatus for providing a
bitstream representing a multi-channel audio signal.
The apparatus comprises a downmixer configured to provide a downmix signal on the
basis of a plurality of audio object signals. The apparatus also comprises a side
information provider configured to provide an object-related parametric side
information, describing characteristics of the audio object signals and downmix
parameters, and a linear combination parameter describing contributions of a user-
specified rendering matrix and of a target rendering matrix to a modified rendering
matrix. The apparatus for providing a bitstream also comprises a bitstream formatter
configured to provide a bitstream comprising a representation of the downmix signal,
the object-related parametric side information and the linear combination parameter.
This apparatus for providing a bitstream representing a multi-channel audio signal is
well-suited for cooperation with the above-discussed apparatus for providing an upmix
signal representation. The apparatus for providing a bitstream representing a multi-
channel audio signal allows for providing the linear combination parameter in
dependence on its knowledge of the audio object signals. Accordingly, the audio
encoder (i.e., the apparatus for providing a bitstream representing a multi-channel audio
signal) can have a strong impact on the rendering quality provided by an audio decoder
(i.e., the above-discussed apparatus for providing an upmix signal representation) which
evaluates the linear combination parameter. Thus, the apparatus for providing the
bitstream representing a multi-channel audio signal has a very high level of control over
the rendering result, which provides for an improved user satisfaction in the many
different scenarios. Accordingly, it is indeed the audio encoder of a service provider
which provides guidance, using the linear combination parameter, whether the user
should be allowed or not to use extreme rendering settings at the risk of audible
distortions. Thus, user disappointment, along with the corresponding negative economic
consequences, can be avoided by using the above-described audio encoder.
Another embodiment according to the invention creates a method for providing an
upmix signal representation on the basis of a downmix signal representation and an
object-related parameter information, which are included in a bitstream representation
of the audio content, in dependence on a user-specified rendering matrix. This method is
based on the same key idea as the above-described apparatus.
Another method according to the invention creates a method for providing a bitstream
representing a multi-channel audio signal. Said method is based on the same finding as
the above-described apparatus.
Another embodiment according to the invention creates a computer program for
performing the above methods.
Another embodiment according to the invention creates a bitstream representing a
multi-channel audio signal. The bitstream comprises a representation of a downmix
signal combining audio signals of a plurality of audio objects in an object-related
parametric side information describing characteristics of the audio objects. The
bitstream also comprises a linear combination parameter describing contributions of a
user-specified rendering matrix and of a target rendering matrix to a modified rendering
matrix. Said bitstream allows for some degree of control over the decoder-sided
rendering parameters from the side of the audio signal encoder.
Brief Description of the Figures
Embodiments according to the present invention will subsequently be described taking
reference to the enclosed figures, in which:
Fig. 1a shows a block schematic diagram of an apparatus for providing an upmix
signal representation, according to an embodiment of the invention;
Fig. lb shows a block schematic diagram of an apparatus for providing a
bitstream representing a multi-channel audio signal, according to an
embodiment of the invention;
Fig. 2 shows a block schematic diagram of an apparatus for providing an upmix
signal representation, according to another embodiment of the invention;
Fig. 3a shows a schematic representation of a bitstream representing a multi-
channel audio signal, according to an embodiment of the invention;
Fig. 3b shows a detailed syntax representation of an SAOC specific
configuration information, according to an embodiment of the invention;
Fig. 3 c shows a detailed syntax representation of an SAOC frame information,
according to an embodiment of the invention;
Fig. 3d shows a schematic representation of an encoding of a distortion control
mode in a bitstream element "bsDcuMode" which can be used in a
SAOC bitstream;
Fig. 3e shows a table representation of an association between a bitstream index
idx and a value of a linear combination parameter "DcuParam[idx]",
which can be used for encoding a linear combination information in an
SAOC bitstream;
Fig 4 shows a block schematic diagram of an apparatus for providing an upmix
signal representation, according to another embodiment of the invention;
Fig, 5a shows a syntax representation of an SAOC specific configuration
information, according to an embodiment of the invention;
Fig. 5b shows a table representation of an association between a bitstream index
idx and a linear combination parameter Param[idx] which can be used for
encoding the linear combination parameter in an SAOC bitstream;
Fig. 6a shows a table describing listening test conditions;
Fig. 6b shows a table describing audio items of the listening tests;
Fig. 6c shows a table describing tested downmix/rendering conditions for a
stereo-to-stereo SAOC decoding scenario;
Fig. 7 shows a graphic representation of distortion control unit (DCU) listening
test results for a stereo-to-stereo SAOC scenario;
Fig. 8 shows a block schematic diagram of a reference MPEG SAOC system;
Fig 9a shows a block schematic diagram of a reference SAOC system using a
separate decoder and mixer;
Fig. 9b shows a block schematic diagram of a reference SAOC system using an
integrated decoder and mixer; and
Fig. 9c shows a block schematic diagram of a reference SAOC system using an
SAOC-to-MPEG transcoder.
Detailed Description of the Embodiments
1. Apparatus for providing an upmix signal representation, according to Fig, la
Fig. la shows a block schematic diagram of an apparatus for providing an upmix signal
representation, according to an embodiment of the invention.
The apparatus 100 is configured to receive a downmix signal representation 110 and an
object-related parametric information 112. The apparatus 100 is also configured to
receive a linear combination parameter 114. The downmix signal representation 110,
the object-related parametric information 112 and the linear combination parameter 114
are all included in a bitstream representation of an audio content. For example, the
linear combination parameter 114 is described by a bitstream element within said
bitstream representation. The apparatus 100 is also configured to receive a rendering
information 120, which defines a user-specified rendering matrix.
The apparatus 100 is configured to provide an upmix signal representation 130, for
example, individual channel signals or an MPEG surround downmix signal in
combination with an MPEG surround side information.
The apparatus 100 comprises a distortion limiter 140 which is configured to obtain a
modified rendering matrix 142 using a linear combination of a user-specified rendering
matrix 144 (which is described, directly or indirectly, by the rendering information 120)
and a target rendering matrix in dependence on a linear combination parameter 146,
which may, for example, be designated with gDCU.
The apparatus 100 may, for example, be configured to evaluate a bitstream element 114
representing the linear combination parameter 146 in order to obtain the linear
combination parameter.
The apparatus 100 also comprises a signal processor 148 which is configured to obtain
the upmix signal representation 130 on the basis of the downmix signal representation
110 and the object-related parametric information 112 using the modified rendering
matrix 142.
Accordingly, the apparatus 100 is capable of providing the upmix signal representation
with good rendering quality using, for example, an SAOC signal processor 148, or any
other object-related signal processor 148. The modified rendering matrix 142 is adapted
by the distortion limiter 140 such that a sufficiently good hearing impression with
sufficiently small distortions is, in most or all cases, achieved. The modified rendering
matrix typically lies "in-between" the user-specified (desired) rendering matrix and the
target rendering matrix, wherein a degree of similarity of the modified rendering matrix
to the user-specified rendering matrix and to the target rendering matrix is determined
by the linear combination parameter, which consequently allows for an adjustment of an
achievable rendering quality and/or of a maximum distortion level of the upmix signal
representation 130.
The signal processor 148 may, for example, be an SAOC signal processor. Accordingly,
the signal processor 148 may be configured to evaluate the object-related parametric
information 112 to obtain parameters describing characteristics of the audio objects
represented, in a downmixed form, by the downmix signal representation 110. In
addition, the signal processor 148 may obtain (for example, receive) parameters
describing the downmix procedure, which is used at the side of an audio encoder
providing the bitstream representation of the audio content in order to derive the
downmix signal representation 110 by combining the audio object signals of a plurality
of audio objects. Thus, the signal processor 148 may, for example, evaluate an object-
level difference information OLD describing a level difference between a plurality of
audio objects for a given audio frame and one or more frequency bands, and an inter-
object correlation information IOC describing a correlation between audio signals of a
plurality of pairs of audio objects for a given audio frame and for one or more frequency
bands. In addition, the signal processor 148 may also evaluate a downmix information
DMG,DCLD describing a downmix, which is performed at the side of an audio encoder
providing the bitstream representation of the audio content, for example, in the form of
one or more downmix gain parameters DMG and one or more downmix channel level
difference parameters DCLD.
In addition, the signal processor 148 receives the modified rendering matrix 142, which
indicates which audio channels of the upmix signal representation 130 should comprise
an audio content of the different audio objects. Accordingly, the signal processor 148 is
configured to determine the contributions of the different audio objects to downmix
signal representation 110 using its knowledge (obtained from the OLD information and
the IOC information) of the audio objects as well as its knowledge of the downmix
process (obtained from the DMG information and the DCLD information).
Furthermore, the signal processor provides the upmix signal representation such that the
modified rendering matrix 142 is considered.
Accordingly, the signal processor 148 fulfills the functionality of the SAOC decoder
820, wherein the downmix signal representation 110 takes the place of the one or more
downmix signals 812, wherein the object-related parametric information 112 takes the
place of the side information 814, and wherein the modified rendering matrix 142 takes
the place of the user interaction/control information 822. The channel signals y1 to
yM take the role of the upmix signal representation 130. Accordingly, reference is made
to the description of the SAOC decoder 820.
Similarly, the signal processor 148 may take the role of the decoder/mixer 920, wherein
the downmix signal representation 110 takes the role of the one or more downmix
signals, wherein the object-related parametric information 112 takes the role of the
object metadata, wherein the modified rendering matrix 142 takes the role of the
rendering information input to the mixer/renderer 926, and wherein the channel signal
928 takes the role of the upmix signal representation 130.
Alternatively, the signal processor 148 may perform the functionality of the integrated
decoder and mixer 950, wherein the downmix signal representation 110 may take the
role of the one or more downmix signals, wherein the object-related parametric
information 112 may take the role of the object metadata, wherein the modified
rendering matrix 142 may take the role of the rendering information input to the object
decoder plus mixer/renderer 950, and wherein the channel signals 958 may take the role
of the upmix signal representation 130.
Alternatively, the signal processor 148 may perform the functionality of the SAOC-to-
MPEG surround transcoder 980, wherein the downmix signal representation 110 may
take the role of the one or more downmix signals, wherein the object-related parametric
information 112 may take the role of the object metadata, wherein the modified
rendering matrix 142 may take the role of the rendering information, and wherein the
one or more downmix signals 988 in combination with the MPEG surround bitstream
984 may take the role of the upmix signal representation 130.
Accordingly, for details regarding the functionality of the signal processor 148,
reference is made to the description of the SAOC decoder 820, of the separate decoder
and mixer 920, of the integrated decoder and mixer 950, and of the SAOC-to-MPEG
surround transcoder 980. Reference is also made, for instance, to documents [3] and [4]
with respect to the functionality of the signal processor 148, wherein the modified
rendering matrix 142, rather than the user-specified rendering matrix 120, takes the role
of the input rendering information in the embodiments according to the invention.
Further details regarding the functionality of the distortion Hmiter 140 will be described
below.
2. Apparatus for providing a bitstream representing a multi-channel audio signal,
according to Fig, lb
Fig. lb shows a block schematic diagram of an apparatus 150 for providing a bitstream
representing a multi-channel audio signal.
The apparatus 150 is configured to receive a plurality of audio object signals 160a to
160N. The apparatus 150 is further configured to provide a bitstream 170 representing
the multi-channel audio signal, which is described by the audio object signals 160a to
160N.
The apparatus 150 comprises a downmixer 180 which is configured to provide a
downmix signal 182 on the basis of the plurality of audio object signals 160a to 160N.
The apparatus 150 also comprises a side information provider 184 which is configured
to provide an object-related parametric side information 186 describing characteristics
of the audio object signals 160a to 160N and downmix parameters used by the
downmixer 180. The side information provider 184 is also configured to provide a
linear combination parameter 188 describing a desired contribution of a (desired) user-
specified rendering matrix and of a target (low-distortion) rendering matrix to a
modified rendering matrix.
The object-related parametric side information 186 may, for example, comprise an
object-level-difference information (OLD) describing object-level-differences of the
audio object signals 160a to 160N (e.g., in a band-wise manner). The object-related
parametric side information may also comprise an inter-object-correlation information
(IOC) describing correlations between the audio object signals 160a to 160N. In
addition, the object-related parametric side information may describe the downmix gain
(e.g., in an object-wise manner), wherein the downmix gain values are used by the
downmixer 180 in order to obtain the downmix signal 182 combining the audio object
signals 160a to 160N. The object-related parametric side information 186 may comprise
a downmix-channel-level-difference information (DCLD), which describes the
differences between the downmix levels for multiple channels of the downmix signal
182 (e.g., if the downmix signal 182 is a multi-channel signal).
The linear combination parameter 188 may for example be a numeric value between 0
and 1, describing to use only a user-specified downmix matrix (e.g., for a parameter
value of 0), only a target rendering matrix (e.g., for a parameter value of 1) or any given
combination of the user-specified rendering matrix and the target rendering matrix in-
between these extremes (e.g., for parameter values between 0 and 1).
The apparatus 150 also comprises a bitstream formatter 190 which is configured to
provide the bitstream 170 such that the bitstream comprises a representation of the
downmix signal 182, the object-related parametric side information 186 and the linear
combination parameter 188.
Accordingly, the apparatus 150 performs the functionality of the SAOC encoder 810
according to Fig. 8 or of the object encoder according to Figs. 9a-9c. The audio object
signals 160a to 160N are equivalent to the object signals x1 to xN received, for example,
by the SAOC encoder 810. The downmix signal 182 may, for example, be equivalent to
one or more downmix signals 812. The object-related parametric side information 186
may, for example, be equivalent to the side information 814 or to the object metadata.
However, in addition to a said 1-channel downmix signal or a multi-channel downmix
signal 182 and said object-related parametric side information 186, the bitstream 170
may also encode the linear combination parameter 188.
Accordingly, the apparatus 150, which can be considered as an audio encoder, has an
impact on a decoder-sided handling of the distortion control scheme, which is
performed by the distortion limiter 140, by appropriately setting the linear combination
parameter 188, such that the apparatus 150 expects a sufficient rendering quality
provided by an audio decoder (e.g. an apparatus 100) receiving the bitstream 170.
For example, the side information provider 184 may set the linear combination
parameter in dependence on a quality requirement information, which is received from
an optional user interface 199 of the apparatus 150. Alternatively, or in addition, the
side information provider 184 may also take into consideration characteristics of the
audio object signals 160a to 160N, and of the downmixing parameters of the
downmixer 180. For example, the apparatus 150 may estimate a degree of distortion,
which is obtained at an audio decoder under the assumption of one or more worst case
user-specified rendering matrices, and may adjust the linear combination parameter 188
such that a rendering quality, which is expected to be obtained by the audio signal
decoder under the consideration of this linear combination parameter, is still considered
as being sufficient by the side information provider 184. For example, the apparatus 150
may set the linear combination parameter 188 to a value allowing for a strong user
impact (influence of the user-specified rendering matrix) onto the modified rendering
matrix, if the side information provider 184 finds that an audio quality of an upmix
signal representation would not be degraded severely even in the presence of extreme
user-specified rendering settings. This may, for example, be the case if the audio object
signals 160a to 160N are sufficiently similar. In contrast, the side information provider
184 may set the linear combination parameter 188 to a value allowing for a
comparatively small impact of the user (or of the user-specified rendering matrix), if the
side information provider 184 finds that extreme rendering settings could lead to strong
audible distortions. This may, for example, be the case if the audio object signals 160a
to 160N are significantly different, such that a clear separation of audio objects at the
side of the audio decoder is difficult (or connected with audible distortions).
It should be noted here that the apparatus 150 may use knowledge for the setting of the
linear combination parameter 188 which is only available at the side to the apparatus
150, but not at the side of an audio decoder (e.g., the apparatus 100), such as, for
example, a desired rendering quality information input to the apparatus 150 via a user
interface or detailed knowledge about the separate audio objects represented by the
audio object signals 160a and 160N.
Accordingly, the side information provider 184 can provide the linear combination
parameter 188 in a very meaningful manner.
3. SAOC System with Distortion Control Unit (DCU). according to Fig. 2
3.1 SAOC Decoder Structure
In the following, a processing performed by a distortion control unit (DCU processing)
will be described taking reference to Fig. 2, which shows a block schematic diagram of
a SAOC system 200. Specifically, Fig. 2 illustrates the distortion control unit DCU
within the overall SAOC system.
Taking reference to Fig. 2, the SAOC decoder 200 is configured to receive a downmix
signal representation 210 representing, for example, a 1-channel downmix signal or a 2-
channel downmix signal, or even a downmix signal having more than two channels. The
SAOC decoder 200 is configured to receive an SAOC bitstream 212, which comprises
an object-related parametric side information, such as, for instance, an object level
difference information OLD, an inter-object correlation information IOC, a downmix
gain information DMG, and, optionally, a downmix channel level difference
information DCLD. The SAOC decoder 200 is also configured to obtain a linear
combination parameter 214, which is also designated with gDCU .
Typically, the downmix signal representation 210, the SAOC bitstream 212 and the
linear combination parameter 214 are included in a bitstream representation of an audio
content.
The SAOC decoder 200 is also configured to receive, for example, from a user
interface, a rendering matrix input 220. For example, the SAOC decoder 200 may
receive a rendering matrix input 220 in the form of a matrix Mren, which defines the
(user-specified, desired) contribution of a plurality of Noijaudio objects to 1,2, or even
more output audio signal channels (of the upmix representation). The rendering matrix
Mren may, for example, be input from a user interface, wherein the user interface may
translate a different user-specified form of representation of a desired rendering setup
into parameters of the rendering matrix Mren. For example, the user-interface may
translate an input in the form of level slider values and an audio object position
information into a user-specified rendering matrix Mren using some mapping.
It should be noted here that throughout the present description, the indices ' defining a
parameter time slot and m defining a processing band are sometimes omitted for the sake
of clarity. Nevertheless, it should be kept in mind that the processing may be performed
individually for a plurality of subsequent parameter time slots having indices 1 and for a
plurality of frequency bands having frequency band indices m.
The SAOC decoder 200 also comprises a distortion control unit DCU 240 which is
configured to receive the user-specified rendering matrix Mren, at least a part of the
SAOC bitstream information 212 (as will be described in detail below) and the linear
combination parameter 214. The distortion control unit 240 provides the modified
rendering matrix Mrenlim .
The audio decoder 200 also comprises an SAOC decoding/transcoding unit 248, which
may be considered as a signal processor, and which receives the downmix signal
representation 210, the SAOC bitstream 212 and the modified rendering matrixMren lim .
The SAOC decoding/transcoding unit 248 provides a representation 230 of one or more
output channels, which may be considered as an upmix signal representation. The
representation 230 of the one or more output channels may, for example, take the form
of a frequency domain representation of individual audio signal channels, of a time
domain representation of individual audio channels or of a parametric multi-channel
representation. For example, the upmix signal representation 230 make take the form of
an MPEG surround representation comprising an MPEG surround downmix signal and
an MPEG surround side information.
It should be noted that the S AOC decoding/transcoding unit 248 may comprise the same
functionality as a signal processor 148, and may be equivalent to the SAOC decoder
820, to the separate coder and mixer 920, to the integrated decoder and mixer 950 and
to the SAOC-to-MPEG surround transcoder 980.
3.2 Introduction into the operation of the SAOC Decoder
In the following, a brief introduction into the operation of the SAOC decoder 200 will
be given.
Within the overall SAOC system, the distortion control unit (DCU) is incorporated into
the SAOC decoder/transcoder processing chain between the rendering interface (e.g., a
user interface at which the user-specified rendering matrix, or an information from
which the user-specified rendering matrix can be derived, is input) and the actual SAOC
decoding/transcoding unit.
The distortion control unit 240 provides a modified rendering matrix MrenHm using the
information from the rendering interface (e.g. the user-specified rendering matrix input,
directly or indirectly, via the rendering interface or user interface) and SAOC data (e.g.,
data from the SAOC bitstream 212). For more details, reference is made to Fig. 2. The
modified rendering matrix Mren lim can be accessed by the application (e.g., the SAOC
decoding/transcoding unit 248), reflecting the actually effective rendering settings.
Based on the user-specified rendering scenario represented by the (user-specified)
rendering matrix M'r^ with elements m'^J, the DCU prevents extreme rendering settings
by producing a modified matrix M';™lim comprising limited rendering coefficients,
which shall be used by the SAOC rendering engine. For all operational modes of
SAOC, the final (DCU processed) rendering coefficients shall be calculated according
to:
The parameter g^y e [0,1], which is also designated as a linear combination parameter,
is used to define the degree of transition from the user specified rendering matrix MJ^J
towards the distortion-free target matrix M'^tar.
The parameter g^y is derived from the bitstream element "bsDcuParam" according to:
Sdcu ~ DcuParam[bsDcuParam].
Accordingly, a linear combination between the user-specified rendering matrix Mren and
the distortion-free target rendering matrix Mren^ar is formed in dependence on the linear
combination parameter g^y. The linear combination parameter g^y is derived from a
bitstream element, such that there is no difficult computation of said linear combination
parameter gDCU required (at least at the decoder side). Also, deriving the linear
combination parameter g^y from the bitstream, including the downmix signal
representation 210, the SAOC bitstream 212 and the bitstream element representing the
linear combination parameter, gives an audio signal encoder a chance to partially
control the distortion control mechanism, which is performed at the side of the SAOC
decoder.
There are two possible versions of the distortion-free target matrix M^tar, suited for
different applications. It is controlled by the bitstream element "bsDcuMode":
• ("bsDcuMode" = 0): The "downmix-similar" rendering, where M'r;"tai
corresponds to the energy normalized downmix matrix.
• ("bsDcuMode" = 1); The "best effort" rendering, where M^jtar is defined as a
function of both downmix and user-specified rendering matrix.
To summarize, there are two distortion control modes called "downmix-similar"
rendering and "best effort" rendering, which can be selected in accordance with the
bitstream elements "bsDcuMode". These two modes differ in the way their target
rendering matrix is computed. In the following, details regarding the computation of the
target rendering matrix for the two modes "downmix-similar" rendering and "best
effort" rendering will be described in detail.
3.3 "Do wnmix-Similar" Rendering
3.3.1 Introduction
The "downmix-similar" rendering method can typically be used in cases where the
downmix is an important reference of artistic high quality. The "downmix-similar"
rendering matrix M^g is computed as
where N1^ represents an energy normalization scalar (for each parameter slot /) and
D'DS is the downmix matrix D' extended by rows of zero elements such that number and
order of the rows of D'DS correspond to the constellation of M;r^,
For example, in the SAOC stereo to multichannel transcoding mode NMPS = 6.
Accordingly D'as is of size NMPS x N (where N depicts the number of input audio
objects) and its rows representing the front left and right output channels equal D' (or
corresponding rows of D').
To facilitate the understanding of the above, the following definitions of the rendering
matrix and of the downmix matrix should be considered.
The (modified) rendering matrix Mren,iim applied to the input audio objects S determines
the target rendered output as Y = Mren,iim S. The (modified) rendering matrix Mren,iim
with elements mjJ maps all input objects /' (i.e., input objects having object index i) to
the desired output channels j (i.e., output channels having channel index j) . The
(modified) rendering matrix Mren,iim is given by
for mono output configuration.
The same dimensions typically also apply to the user-specified rendering matrix Mren
and the target rendering matrix Mren,tar.
The downmix matrix D applied to the input audio objects S (in an audio decoder)
determines the downmix signal as X = DS.
For the stereo downmix case, the downmix matrix D of size 2xN (also designated
with D', to show a possible time dependency) with elements dtJ (i = 0,1; j = 0,..., N -1)
is obtained (in an audio decoder) from the DMG and DCLD parameters as
For the mono downmix case the downmix matrix D of size 1 x N with elements dfJ
(i = 0;;' = 0„.., AT -1) is obtained (in an audio decoder) from the DMG parameters as
The downmix parameters DMG and DCLD are obtained from the SAOC bitstream 212.
3.3.2 Computation of the Energy Normalization Scalar for all decoding/transcoding
SAOC modes
For all decoding/transcoding SAOC modes the energy normalization scalar N'DS is
computed using the following equation:
3.4 "Best-Effort" Rendering
3.4.1 Introduction
The "best effort" rendering method can typically be used in cases where the target
rendering is an important reference.
The "best effort" rendering matrix describes a target rendering matrix, which depends
on the downmix and rendering information. The energy normalization is represented by
a matrix Ng" of size NMPS x M, hence it provides individual values for each output
channel. This requires different calculations of N'B*£ for the different SAOC operation
modes, which are outlined in the following. The "best effort" rendering matrix is
computed as
Here D' is the downmix matrix and N^ represents the energy normalization matrix.
The square root operator in the above equation designates an element-wise square root
formation.
In the following, the computation of the value N'B£, which may be an energy
normalization scalar in the case of an SAOC mono-to-mono decoding mode, and which
may be an energy normalization matrix in the case of other decoding modes or
transcoding modes, will be discussed in detail.
3.4.2 SAOC mono-to-mono ("x-1-1") decoding mode
For the "x-1-1" SAOC mode in which a mono downmix signal is decoded to obtain a
mono output signal (as an upmix signal representation), the energy normalization scalar
N'^ is computed using the following equation
3.4.3 SAOC mono-to-stereo ("x-1-2") decoding mode
For the "x-1-2" SAOC mode, in which a mono downmix signal is decoded to obtain a
stereo (2-channel) output (as an upmix signal representation), the energy normalization
matrix N'^ of size 2 x 1 is computed using the following equation
3.4.4 SAOC mono-to-binaural ("x-l-b") decoding mode
For the "x-l-b" SAOC mode, in which a mono downmix signal is decoded to obtain a
binaural rendered output signal (as an upmix signal representation), the energy
normalization matrix Ng™ of size 2 x 1 is computed using the following equation
f N-\ , N-l , \T
SJ.m ( J,m\ , „ V"* J,m I J,m\ , „
l^BE N-l , > JV-1 ,
sw)+« nth*
The elements a'^y comprise (or are taken from) the target binaural rendering
matrix A',m.
3.4.5 SAOC stereo-to-mono ("x-2-1") decoding mode
For the "x-2-1" SAOC mode, in which a two-channel (stereo) downmix signal is
decoded to obtain a one-channel (mono) output signal (as an upmix signal
representation), the energy normalization matrix N'j™ of size 1x2 is computed using
the following equation
where M'^ is mono rendering matrix of size 1 x JV.
3.4.6 SAOC stereo-to-stereo ("x-2-2") decoding mode
For the "x-2-2" SAOC mode, in which a stereo downmix signal is decoded to obtain a
stereo output signal (as an upmix signal representation), the energy normalization
matrix N'B*£ of size 2x2 is computed using the following equation
where M';" is stereo rendering matrix of size 2 x JV.
3.4.7 SAOC stereo-to-binaural ("x-2-b") decoding mode
For the "x-2-b" SAOC mode, in which a stereo downmix signal is decoded to obtain a
binaural-rendered output signal (as an upmix signal representation), the energy
normalization matrix N'fl*£ of size 2 x 2 is computed using the following equation
where A'ffl is a binaural rendering matrix of size 2 x N.
3.4.8 SAOC mono-to-multichannel ("x-1-5") transcoding mode
For the "x-1-5" SAOC mode, in which a mono downmix signal is transcoded to obtain a
5-channel or 6-channel output signal (as an upmix signal representation), the energy
normalization matrix Ng™ of size NMI,S xl is computed using the following equation
3.4.9 SAOC stereo-to-multichannel ("x-2-5") transcoding mode
For the "x-2-5" SAOC mode, in which a stereo downmix signal is transcoded to obtain
a 5-channel or 6-channel output signal (as an upmix signal representation), the energy
normalization matrix N'^™ of size NMPS x 2 is computed using the following equation
3.4.10 Computation of J'
To avoid numerical problems when calculating the term J' = ID' (d') I in 3.4.5, 3.4.6,
3.4.7, and 3.4.9, J' is modified in some embodiments. First the eigenvalues \2 of J'
are calculated, solving det( J - \ 2I) = 0.
Eigenvalues are sorted in descending (\ > /^) order and the eigenvector corresponding
to the larger eigenvalue is calculated according to the equation above. It is assured to lie
in the positive x-plane (first element has to be positive). The second eigenvector is
obtained from the first by a - 90 degrees rotation:
3.4.11 Distortion Control Unit (DCU) application for enhanced audio objects (EAO)
In the following, some optional extensions regarding the application of the distortion
control unit will be described, which may be implemented in some embodiments
according to the invention.
For SAOC decoders that decode residual coding data and thus support the handling of
EAOs, it can be meaningful to provide a second parameterization of the DCU which
allows taking advantage of the enhanced audio quality provided by the use of EAOs.
This is achieved by decoding and using a second alternate set of DCU parameters (i.e.
bsDcuMode2 and bsDcuParam2) which is additionally transmitted as part of the data
structures containing residual data (i.e. SAOCExtensionConfigData() and
SAOCExtensionFrameData()). An application can make use of this second parameter
set if it decodes residual coding data and operates in strict EAO mode which is defined
by the condition that only EAOs can be modified arbitrarily while all non-EAOs only
undergo a single common modification. Specifically, this strict EAO mode requires
fulfillment of two following conditions:
The downmix matrix and rendering matrix have the same dimensions (implying that the
number of rendering channels is equal to the number of downmix channels).
The application only employs rendering coefficients for each of the regular objects (i.e.
non-EAOs) that are related to their corresponding downmix coefficients by a single
common scaling factor.
4. Bitstream according to Fig. 3a
In the following, a bitstream representing a multi-channel audio signal will be described
taking reference to Fig. 3a which shows a graphical representation of such a bitstream
300.
The bitstream 300 comprises a downmix signal representation 302, which is a
representation (e.g., an encoded representation) of a downmix signal combining audio
signals of a plurality of audio objects. The bitstream 300 also comprises an object-
related parametric side information 304 describing characteristics of the audio object
and, typically, also characteristics of a downmix performed in an audio encoder. The
object-related parametric information 304 preferably comprises an object level
difference information OLD, an inter-object correlation information IOC, a downmix
gain information DMG and a downmix channel level different information DCLD. The
bitstream 300 also comprises a linear combination parameter 306 describing desired
contributions of a user-specified rendering matrix and of a target rendering matrix to a
modified rendering matrix (to be applied by an audio signal decoder).
Further optional details regarding this bitstream 300, which may be provided by the
apparatus 150 as the bitstream 170, and which may be input into the apparatus 100 to
obtain the downmix signal representation 110, the object-related parametric information
112 and the linear combination parameter 140, or into the apparatus 200 to obtain the
downmix information 210, the SAOC bitstream information 212 and the linear
combination parameter 214, will be described in the following taking reference to Figs.
3b and 3c.
5. Bitstream Syntax Details
5.1. SAOC Specific Configuration Syntax
Fig. 3b shows a detailed syntax representation of an SAOC specific configuration
information.
The SAOC specific configuration 310 according to Fig. 3b may, for example, be part of
a header of the bitstream 300 according to Fig. 3a.
The SAOC specific configuration may, for example, comprise a sampling frequency
configuration describing a sampling frequency to be applied by an SAOC decoder. The
SAOC specific configuration also comprises a low-delay-mode configuration describing
whether a low-delay mode or a high-delay mode of the signal processor 148 or of the
SAOC decoding/transcoding unit 248 should be used. The SAOC specific configuration
also comprises a frequency resolution configuration describing a frequency resolution to
be used by the signal processor 148 or by the SAOC decoding/transcoding unit 248. In
addition, the SAOC specific configuration may comprise a frame length configuration
describing a length of audio frames to be used by the signal processor 148, or by the
SAOC decoding/transcoding unit 248. Moreover, the SAOC specific configuration
typically comprises an object number configuration describing a number of audio
objects to be processed by the signal processor 148, or by the SAOC
decoding/transcoding unit 248. The object number configuration also describes a
number of object-related parameters included in the object-related parametric
information 112, or in the SAOC bitstream 212. The SAOC specific configuration may
comprise an object-relationship configuration, which designates objects having a
common object-related parametric information. The SAOC specific configuration may
also comprise an absolute energy transmission configuration, which indicates whether
an absolute energy information is transmitted from an audio encoder to an audio
decoder. The SAOC specific configuration may also comprise a downmix channel
number configuration, which indicates whether there is only one downmix channel,
whether there are two downmix channels, or whether there are, optionally, more than
two downmix channels. In addition, the SAOC specific configuration may comprise
additional configuration information in some embodiments.
The SAOC specific configuration may also comprise post-processing downmix gain
configuration information "bsPdgFlag" which defines whether a post processing
downmix gain for an optional post-processing are transmitted.
The SAOC specific configuration also comprises a flag "bsDcuFlag" (which may, for
example, be a 1-bit flag), which defines whether the values "bsDcuMode" and
"bsDcuParam" are transmitted in the bitstream. If this flag "bsDcuFlag" takes the value
of i", another flag which is marked "bsDcuMandatory" and a flag "bsDcuDynamic"
are included in the SAOC specific configuration 310. The flag "bsDcuMandatory"
describes whether the distortion control must be applied by an audio decoder. If the flag
"bsDcuMandatory" is equal to 1, then the distortion control unit must be applied using
the parameters "bsDcuMode" and "bsDcuParam" as transmitted in the bitstream. If the
flag "bsDcuMandatory" is equal to "0", then the distortion control unit parameters
"bsDcuMode" and "bsDcuParam" transmitted in the bitstream are only recommended
values and also other distortion control unit settings could be used.
In other words, an audio encoder may activate the flag "bsDcuMandatory"in order to
enforce the usage of the distortion control mechanism in a standard-compliant audio
decoder, and may deactivate said flag in order to leave the decision whether to apply the
distortion control unit, and if so, which parameters to use for the distortion control unit,
to the audio decoder.
The flag "bsDcuDynamic" enables a dynamic signaling of the values "bsDcuMode" and
"bsDcuParam". If the flag "bsDcuDynamic" is deactivated, the parameters
"bsDcuMode" and "bsDcuParam" are included in the SAOC specific configuration, and
otherwise, the parameters "bsDcuMode" and "bsDcuParam" are included in the SAOC
frames, or, at least, in some of the SAOC frames, as will be discussed later on.
Accordingly, an audio signal encoder can switch between a one-time signaling (per
piece of audio comprising a single SAOC specific configuration and, typically, a
plurality of SAOC frames) and a dynamic transmission of said parameters within some
or all of the SAOC frames.
The parameter "bsDcuMode" defines the distortion-free target matrix type for the
distortion control unit (DCU) according to the table of Fig. 3d.
The parameter "bsDcuParam" defines the parameter value for the distortion control unit
(DCU) algorithm according to the table of Fig. 3e. In other words, the 4-bit parameter
"bsDcuParam" defines an index value idx, which can be mapped by an audio signal
decoder onto a linear combination value g^y (also designated with "DcuParamfind]"
or "DcuParam[idx]"). Thus, the parameter "bsDcuParam" represents, in a quantized
manner, the linear combination parameter.
As can be seen in Fig. 3b, the parameters "bsDcuMandatory", "bsDcuDynamic",
"bsDcuMode" and "bsDcuParam" are set to a default value of "0", if the flag
"bsDcuFlag" takes the value of "0", which indicates that no distortion control unit
parameters are transmitted.
The SAOC specific configuration also comprises, optionally, one or more byte
alignment bits "ByteAlign()" to bring the SAOC specific configuration to a desired
length.
In addition, the SAOC specific configuration may optionally comprise a SAOC
extension configuration "SAOCExtensionConfigO", which comprises additional
configuration parameters. However, said configuration parameters are not relevant for
the present invention, such that a discussion is omitted here for the sake of brevity.
5.2. SAOC Frame Syntax
In the following the syntax of an SAOC frame will be described taking reference to Fig.
3c.
The SAOC frame "SAOCFrame" typically comprises encoded object level difference
values OLD as discussed before, which may be included in the SAOC frame data for a
plurality of frequency bands ("band-wise") and for a plurality of audio objects (per
audio object).
The SAOC frame also, optionally, comprises encoded absolute energy values NRG
which may be included for a plurality of frequency bands (band-wise).
The SAOC frame may also comprise encoded inter-object correlation values IOC,
which are included in the SAOC frame data for a plurality of combinations of audio
objects. The IOC values are typically included in a band-wise manner.
The SAOC frame also comprises encoded downmix-gain values DMG, wherein there is
typically one downmix gain value per audio object per SAOC frame.
The SAOC frame also comprises, optionally, encoded downmix channel level
differences DCLD, wherein there is typically one downmix channel level difference
value per audio object and per SAOC frame.
Also, the SAOC frame typically comprises, optionally, encoded post-processing
downmix gain values PDG.
In addition, an SAOC frame may also comprise, under some circumstances, one or
more distortion control parameters. If the flag "bsDcuFlag", which is included in the
SAOC specific configuration section, is equal to "1", indicating usage of distortion
control unit information in the bitstream, and if the flag "bsDcuDynamic" in the SAOC
specific configuration also takes the value of "1", indicating the usage of a dynamic
(frame-wise) distortion control unit information, the distortion control information is
included in the SAOC frame, provided that the SAOC frame is a so-called
"independent" SAOC frame, for which the flag "bsIndependencyFlag" is active or that
the flag "bsDcuDynamicUpdate" is active.
It should be noted here that the flag "bsDcuDynamicUpdate" is only included in the
SAOC frame if the flag "bsIndependencyFlag" is inactive and that the flag
"bsDcuDynamicUpdate" defines whether the values "bsDcuMode" and "bsDcuParam"
are updated. More precisely, "bsDcuDynamicUpdate" = = 1 means that the values
"bsDcuMode" and "bsDcuParam" are updated in the current frame, whereas
"bsDcuDynamicUpdate" = = 0 means that the previously transmitted values are kept.
Accordingly, the parameters "bsDcuMode" and "bsDcuParam", which have been
explained above, are included in the SAOC frame if the transmission of distortion
control unit parameters is activated and a dynamic transmission of the distortion control
unit data is also activated and the flag "bsDcuDynamicUpdate" is activated. In addition,
the parameters "bsDcuMode" and "bsDcuParam" are also included in the SAOC frame
if the SAOC frame is an "independent" SAOC frame, the transmission of distortion
control unit data is activated and the dynamic transmission of distortion control unit
data is also activated.
The SAOC frame also comprises, optionally, fill data "byteAlign()" to fill up the SAOC
frame to a desired length.
Optionally, the SAOC frame may comprise additional information, which is designated
as "SAOCExt or ExtensionFrame()". However, this optional additional SAOC frame
information is not relevant for the present invention and, for the sake of brevity, will
therefore not be discussed here.
For completeness, it should be noted that the flag "bsIndependencyFlag" indicates if
lossless coding of the current SAOC frame is done independently of the previous SAOC
frame, i.e. whether the current SAOC frame can be decoded without knowledge of the
previous SAOC frame.
6. SAOC decoder/transcoder according to Fig. 4
In the following, further embodiments of rendering coefficient limiting schemes for
distortion control in SAOC will be described.
6.1 Overview
Fig. 4 shows a block schematic diagram of an audio decoder 400, according to an
embodiment of the invention.
The audio decoder 400 is configured to receive a downmix signal 410, an SAOC
bitstream 412, a linear combination parameter 414 (also designated with A), and a
rendering matrix information 420 (also designated with R). The audio decoder 400 is
configured to receive an upmix signal representation, for example, in the form of a
plurality of output channels 130a to 130M. The audio decoder 400 comprises a
distortion control unit 440 (also designated with DCU) which receives at least a part of
the SAOC bitstream information of the SAOC bitstream 412, the linear combination
parameter 414 and the rendering matrix information 420. The distortion control unit
provides a modified rendering information Rijm which may be a modified rendering
matrix information.
The audio decoder 400 also comprises an SAOC decoder and/or SAOC transcoder 448,
which receives the downmix signal 410, the SAOC bitstream 412 and the modified
rendering information Riim and provides, on the basis thereof, the output channels 130a
to 130M.
In the following, the functionality of the audio decoder 400, which uses one or more
rendering coefficient limiting schemes according to the present invention, will be
discussed in detail.
The general SAOC processing is carried out in a time/frequency selective way and can
be described as follows. The SAOC encoder (for example, the SAOC encoder 150)
extracts the psychoacoustic characteristics (e.g. object power relations and correlations)
of several input audio object signals and then downmixes them into a combined mono
or stereo channel (for example, the downmix signal 182 or the downmix signal 410).
This downmix signal and extracted side information (for example, the object-related
parametric side information or the SAOC bitstream information 412 are transmitted (or
stored) in compressed format using the well-known perceptual audio coders. On the
receiving end, the SAOC decoder 418 conceptually tries to restore the original object
signals (i.e. separate downmixed objects) using the transmitted side information 412.
These approximated object signals are then mixed into a target scene using a rendering
matrix. The rendering matrix for example R or R|jm is composed of the Rendering
Coefficients (RCs) specified for each transmitted audio object and upmix setup
loudspeaker. These RCs determine gains and spatial positions of all separated/rendered
objects.
Effectively, the separation of the object signals is rarely or even never executed since
the separation and the mixing is performed in a single combined processing step which
results in an enormous reduction of computational complexity. This scheme is
tremendously efficient, both in terms of transmission bitrate (only needs to transmit one
or two downmix channels 182, 410 plus some side information 186, 188, 412, 414,
instead of a number of individual object audio signals) and computational complexity
(the processing complexity relates mainly to the number of output channels rather than
the number of audio objects). The SAOC decoder transforms (on a parametric level) the
object gains and other side information directly into the Transcoding Coefficients (TCs)
which are applied to the downmix signal 182, 414 to create the corresponding signals
130a to 130M for the rendered output audio scene (or preprocessed downmix signal for
a further decoding operation, i.e. typically multichannel MPEG Surround rendering).
The subjectively perceived audio quality of the rendered output scene can be improved
by application of a distortion control unit DCU (e.g. a rendering matrix modifying unit),
as described in [6]. This improvement can be achieved for the price of accepting a
moderate dynamic modification of the target rendering settings. The modification of the
rendering information can be done time and frequency variant, which under specific
circumstances may result in unnatural sound colorations and/or temporal fluctuation
artifacts.
Within the overall SAOC system, the DCU can be incorporated into the SAOC
decoder/transcoder processing chain in the straightforward way. Namely, it is placed at
the front-end of the SAOC by controlling the RCs R, see Fig. 4.
6.2 Underlying hypothesis
The underlying hypothesis of the indirect control method considers a relationship
between distortion level and deviations of the RCs from their corresponding objects'
level in the downmix. This is based on the observation that the more specific
attenuation/boosting is applied by the RCs to a particular object with respect to the other
objects, the more aggressive modification of the transmitted downmix signal is to be
performed by the SAOC decoder/transcoder. In other words: the higher the deviation of
the "object gain" values are relative to each other, the higher the chance for
unacceptable distortion to occur (assuming identical downmix coefficients).
6.3 Calculation of the limited rendering coefficients
Based on the user specified rendering scenario represented by the coefficients (the RCs)
of a matrix R of size ch x ob (i.e. the rows correspond to the output channels 130a to
130M, the columns to the input audio objects), the DCU prevents extreme rendering
D
settings by producing a modified matrix lim comprising limited rendering coefficients,
which are actually used by the SAOC rendering engine 448. Without loss of generality,
in the subsequent description the RCs are assumed to be frequency invariant to simplify
the notation. For all operational modes of SAOC the limited rendering coefficients can
be derived as
This means that by incorporating the cross-fading parameter A 6 [0,1] (also designated
as a linear combination parameter), a blending of the (user specified) rendering matrix
R towards a target matrix R can be realized. In other words, the limited matrix Rlim
represents a linear combination of the rendering matrix R and a target matrix. On one
hand, the target rendering matrix could be the downmix matrix (i.e. the downmix
channels are passed through the transcoder 448) with a normalization factor or another
static matrix that results in a static transcoding matrix. This "downmix-similar
rendering" ensures that the target rendering matrix does not introduce any SAOC
processing artifacts and consequently represents an optimal rendering point in terms of
audio quality albeit being totally regardless of the initial rendering coefficients.
However, if an application demands a specific rendering scenario or a user set high
value on his/her initial rendering setup (especially, for example, the spatial position of
one or more objects), the downmix-similar rendering fails to serve as target point. On
the other hand, such a point can be interpreted as "best-effort rendering" when taking
into account both the downmix and the initial rendering coefficients (for example, the
user specified rendering matrix). The aim of this second definition of the target
rendering matrix is to preserve the specified rendering scenario (for example, defined by
the user-specified rendering matrix) in a best possible way, but at the same time keeping
the audible degradation due to excessive object manipulation on a minimum level.
6.4 Downmix Similar rendering
6.4.1 Introduction
The downmix matrix D of size Ndmx x Nob is determined by the encoder (for example,
the audio encoder 150) and comprises information on how the input objects are linearly
combined into the downmix signal which is transmitted to the decoder. For example,
with a mono downmix signal, D reduces to a single row vector, and in the stereo
downmix case Ndmx = 2.
The "downmix-similar rendering" matrix R^ is computed as
R(=RDS) = NDSDR,
where NDS represents the energy normalization scalar and DR is the downmix matrix
extended by rows of zero elements such that number and order of the rows of DR
correspond to the constellation of R . For example, in the SAOC stereo to multichannel
transcoding mode (x-2-5) Ndmx = 2 and Nch = 6. Accordingly DR is of size Nch x Nob
and its rows representing the front left and right output channels equal D.
6.4.2 All decoding/transcoding SAOC modes
For all decoding/transcoding SAOC modes the energy normalization scalar NDS can be
computed using the following equation
where the operator trace(X) implies summation of all diagonal elements of matrix X,
The (*) implies the complex conjugate transpose operator.
6.5 Best effort rendering
6.5.1 Introduction
The best effort rendering method describes a target rendering matrix, which depends on
the downmix and rendering information. The energy normalization is represented by a
matrix NBE of size N^ x Ndmx, hence it provides individual values for each output
channel (provided that there is more than one output channel). This requires different
calculations of NBE for the different SAOC operation modes, which are outlined in the
subsequent sections.
The "best effort rendering" matrix is computed as
where D is the downmix matrix and NBE represents the energy normalization matrix.
6.5.2 SAOC mono-to-mono ("x-1 -1") decoding mode
For the "x-1-1" SAOC mode the energy normalization scalar NBE can be computed
using the following equation
6.5.3 SAOC mono-to-stereo ("x-1-2") decoding mode
For the "x-1-2" SAOC mode the energy normalization matrix NBH of size 2x1 can be
computed using the following equation
6.5.4 SAOC mono-to-binaural ("x-l-b") decoding mode
For the "x-l-b" SAOC mode the energy normalization matrix NBE of size 2x1 can be
computed using the following equation
It should be noted further that here ri and t2 consider/incorporate binaural HRTF
parameter information.
It should also be noted that for all 3 equations above, the square root of Nbe must be
taken, i.e.
(see description before).
6.5.5 SAOC stereo-to-mono ("x-2-1") decoding mode
For the "x-2-1" SAOC mode the energy normalization matrix NBE of size 1x2 can be
computed using the following equation
where the mono rendering matrix R: of size 1 x Nob is defined as
6.5.6 SAOC stereo-to-stereo ("x-2-2") decoding mode
For the "x-2-2" SAOC mode the energy normalization matrix NBE of size 2x2 can be
computed using the following equation
where the stereo rendering matrix R2 of size 2xNob is defined as
6.5.7 SAOC mono-to-binaural ("x-2-b") decoding mode
For the "x-2-b" SAOC mode the energy normalization matrix NBB of size 2x2 can be
computed using the following equation
where the binaural rendering matrix R2 of size 2xNob is defined as
It should be noted further that here r\fl and V2,n consider/incorporate binaural HRTF
parameter information.
6.5.8 SAOC mono-to-multichannel ("x-1-5") transcoding mode
For the "x-1-5" SAOC mode the energy normalization matrix NSE of size Nch xl can
be computed using the following equation
Again, taking the square-root for each element is recommended or even required in
some cases.
6.5.9 SAOC stereo-to-multichannel ("x-2-5") transcoding mode
For the "x-2-5" SAOC mode the energy normalization matrix NBE of size N^ x 2 can
be computed using the following equation
6.5.10 Computation of the (DD*)"1
For the computation of the term (DD* ) regularization methods can be applied to
prevent ill-posed matrix results.
6.6 Control of the rendering coefficient limiting schemes
6.6.1 Example of bitstream syntax
In the following a syntax representation of a SAOC specific configuration will be
described taking reference to Fig. 5 a. The SAOC specific configuration
"SAOCSpecificConfigO" comprises conventional SAOC configuration information.
Moreover, the SAOC specific configuration comprises a DCU specific addition 510,
which will be described in more detail in the following. The SAOC specific
configuration also comprises one or more fill bits "ByteAlign()", which may be used to
adjust the length of the SAOC specific configuration. In addition, the SAOC specific
configuration may optionally comprise and SAOC extension configuration, which
comprises further configuration parameters.
The DCU specific addition 510 according to Fig. 5 a to the bitstream syntax element
"SAOCSpecificConfigO" is an example of bitstream signaling for the proposed DCU
scheme. This relates to the syntax described in sub-clause "5.1 pay loads for SAOC" of
the draft SAOC Standard according to reference [8].
In the following, the definition of some of the parameters will be given.
"bsDcuFlag" Defines whether the settings for the DCU are determined by the
SAOC encoder or decoder/transcoder. More precisely,
"bsDcuFlag" = 1 means that the values "bsDcuMode" and
"bsDcuParam" specified in the SAOCSpecificConfigO by the
SAOC encoder are applied to the DCU, whereas "bsDcuFlag"
= 0 means that the variables "bsDcuMode" and "bsDcuParam"
(initialized by the default values) can be further modified by the
SAOC decoder/transcoder application or user.
"bsDcuMode" Defines the mode of the DCU. More precisely, "bsDcuMod" =
0 means that the "downmix-similar" rendering mode is applied
by the DCU, whereas "bsDcuMode" = 1 that the "best-effort"
rendering mode is applied by the DCU algorithm.
"bsDcuParam" Defines the blending parameter value for the DCU algorithm,
wherein the table of Fig. 5b shows a quantization table for the
"bsDcuParam" parameters.
The possible "bsDcuParam" values are in this example part of a table with 16 entries
represented by 4 bits. Of course any table, bigger or smaller, could be used. The spacing
between the values can be logarithmic in order to correspond to maximum object
separation in decibels. But the values could also be linearly spaced, or a hybrid
combination of logarithmic and linear, or any other kind of scale.
The "bsDcuMode" parameter in the bitstream makes it possible for at the encoder side
choosing an, for the situation, optimal DCU algorithm. This can be very useful since
some applications or content might benefit from the "downmix-similar" rendering mode
while other might benefit from the "best-effort" rendering mode.
Typically, the "downmix-similar" rendering mode can be the desired method for
applications where backward/forward compatibility is important and the downmix has
important artistic qualities that needs to be preserved. On the other hand, the "best-
effort" rendering mode can have better performance in cases where this is not the case.
These DCU parameters related to the present invention could of course be conveyed in
any other parts of the SAOC bitstream. An alternative location would be using the
"SAOCExtensionConfigO" container where a certain extension ID could be used. Both
these sections are located in the SAOC header, assuring minimum data-rate overhead.
Another alternative is to convey the DCU data in the payload data (i.e. in
SAOCFrame()), This would allow for time-variant signaling (for example, signal
adaptive control).
A flexible approach is to define bitstream signaling of the DCU data for both header
(i.e. static signaling) and in the payload data (i.e. dynamic signaling). Then an SAOC
encoder is free to choose one of the two signaling methods.
6.7 Processing Strategy
In the case if the DCU settings (e.g. DCU mode "bsDcuMode" and blending parameter
setting "bsDcuParam") are explicitly specified by the SAOC encoder (e.g.
"bsDcuFlag"=l), the SAOC decoder/transcoder applies these values directly to the
DCU. If the DCU settings are not explicitly specified (e.g. "bsDcuFlag"=0) the SAOC
decoder/transcoder uses the default values and allows the SAOC decoder/transcoder
application or user to modify them. The first quantization index (e.g. idx=0) can be used
for disabling DCU. Alternatively, the DCU default value ("bsDcuParam") can be "0"
i.e. disabling the DCU or "1" i.e. full limiting.
7. Performance Evaluation
7.1 Listening test design
A subjective listening test has been conducted to assess the perceptual performance of
the proposed DCM concept and compare it to the results of the regular SAOC RM
decoding/transcoding processing. Compared to other listening tests, the task of this test
is to consider best possible reproduction quality in extreme rendering situations
("soloing objects", "muting objects") regarding two quality aspects:
1. achieving the objective of the rendering (good attenuation/boosting of the target
objects)
2. overall scene sound quality (considering distortions, artifacts, unnaturalness...)
Please note that an unmodified SAOC processing may fulfill aspect #1 but not aspect
#2, whereas simply using the transmitted downmix signal may fulfill aspect #2 but not
aspect #1.
The listening test was conducted presenting only true choices to the listener, i.e. only
material that is truly available as a signal at the decoder side. Thus, the presented signals
are the output signal of the regular (unprocessed by the DCU) SAOC decoder,
demonstrating the baseline performance of the SAOC and the SAOC/DCU output. In
addition, the case of trivial rendering, which corresponds to the downmix signal, is
presented in the listening test.
The table of Fig. 6a describes the listening test conditions.
Since the proposed DCU operates using the regular SAOC data and downmixes and
does not rely on residual information, no core coder has been applied to the
corresponding SAOC downmix signals.
7.2 Listening test items
The following items together with extreme and critical rendering have been chosen for
the current listening test from the CfP listening test material.
The table of Fig. 6b describes the audio items of the listening tests.
7.3 Downmix and rendering settings
The rendering objects gains which are described in the table of Fig. 6c have been
applied for the considered upmix scenarios.
7.4 Listening test instructions
The subjective listening tests were conducted in an acoustically isolated listening room
that is designed to permit high-quality listening. The playback was done using
headphones (STAX SR Lambda Pro with Lake-People D/A-Converter and ST AX
SRM-Monitor).
The test method followed the procedure used in the spatial audio verification tests,
similar to the "Multiple Stimulus with Hidden Reference and Anchors" (MUSHRA)
method for the subjective assessment of intermediate quality audio [2]. The test method
has been modified as described above in order to assess the perceptual performance of
the proposed DCU. The listeners were instructed to adhere to the following listening
test instructions:
"Application scenario: Imagine you are the user of an interactive music remix system
which allows you to make dedicated remixes of music material. The system provides
mixing desk style sliders for each instrument to change its level, spatial position, etc.
Due to the nature of the system, some extreme sound mixes can lead to distortion which
degrades the overall sound quality. On the other hand, sound mixes with similar
instrument levels tend to produce better sound quality.
It is the objective of this test to assess different processing algorithms regarding their
impact on sound modification strength and sound quality.
There is no "Reference signal" in this test! Instead of that a description of the desired
sound mixes is given below.
For each audio item please:
first read the description of the desired sound mixes that you as a system user
would like to achieve
Item "BlackCoffee": Soft brass section within the sound mix
Item "VoiceOverMusic": Soft background music
Item "Audition": Strong vocal sound and soft music
Item "LovePop": Soft string section within the sound mix
then grade the signals using one common grade to describe both
achieving the rendering objective of the desired sound mix
overall scene sound quality (consider distortions, artifacts, unnaturalness,
spatial distortions,...)"
A total of 8 listeners participated in each of the performed tests. All subjects can be
considered as experienced listeners. The test conditions were randomized automatically
for each test item and for each listener. The subjective responses were recorded by a
computer-based listening test program on a scale ranging from 0 to 100, with five
intervals labeled in the same way as on the MUSHRA scale. An instantaneous
switching between the items under test was allowed.
7.5 Listening test results
The plots shown in the graphical representation of Fig. 7 show the average score per
item over all listeners and the statistical mean value over all evaluated items together
with the associated 95% confidence intervals.
The following observations can be made based upon the results of the conducted
listening tests: For conducted listening test the obtained MUSHRA scores prove that the
proposed DCU functionality provides a significantly better performance in comparison
with the regular SAOC RM system in sense of overall statistical mean values. One
should note that the quality of all items produced by the regular SAOC decoder
(showing strong audio artifacts for the considered extreme rendering conditions) is
graded as low as the quality of downmix-identical rendering settings which does not
fulfill the desired rendering scenario at all. Hence, it can be concluded that the proposed
DCU methods lead to considerable improvement of subjective signal quality for all
considered listening test scenarios.
8. Conclusions
To summarize the above discussion, rendering coefficient limiting schemes for
distortion control in SAOC have been described. Embodiments according to the
invention may be used in combination with parametric techniques for bitrate-efficient
transmission/storage of audio scenes containing multiple audio objects, which have
recently been proposed (e.g., see references [1], [2], [3], [4] and [5]).
In combination with user interactivity at the receiving side, such techniques may
conventionally (without the use of the inventive rendering coefficient limiting schemes)
lead to a low quality of the output signals if extreme object rendering is performed (see,
for example, reference [6]).
The present specification is focused on Spatial Audio Object Coding (SAOC) which
provides means for a user interface for the selection of the desired playback setup (e.g.
mono, stereo, 5.1, etc.) and interactive real-time modification of the desired output
rendering scene by controlling the rendering matrix according to personal preference or
other criteria. However, the invention is also applicable for parametric techniques in
general.
Due to the downmix/separation/mix-based parametric approach, the subjective quality
of the rendered audio output depends on the rendering parameter settings. The freedom
of selecting rendering settings of the user's choice entails the risk of the user selecting
inappropriate object rendering options, such as extreme gain manipulations of an object
within the overall sound scene.
For a commercial product, it is by all means unacceptable to produce bad sound quality
and/or audio artifacts for any settings on the user interface. In order to control excessive
deterioration of the produced SAOC audio output, several computational measures have
been described which are based on the idea of computing a measure of perceptual
quality of the rendered scene, and depending on this measure (and, optionally, other
information), modify the actually applied rendering coefficients (see, for example,
reference [6]).
The present document describes alternative ideas for safeguarding the subjective sound
quality of the rendered SAOC scene for which all processing is carried out entirely
within the SAOC decoder/transcoder, and which do not involve the explicit calculation
of sophisticated measures of perceived audio quality of the rendered sound scene.
These ideas can thus be implemented in a structurally simple and extremely efficient
way within the SAOC decoder/transcoder framework. The proposed Distortion Control
Unit (DCU) algorithm aims at limiting input parameters of the SAOC decoder, namely,
the rendering coefficients.
To summarize the above, embodiments according to the invention create an audio
encoder, an audio decoder, a method of encoding, a method of decoding, and computer
programs for encoding or decoding, or encoded audio signals as described above.
9. Implementation alternatives
Although some aspects have been described in the context of an apparatus, it is clear
that these aspects also represent a description of the corresponding method, where a
block or device corresponds to a method step or a feature of a method step.
Analogously, aspects described in the context of a method step also represent a
description of a corresponding block or item or feature of a corresponding apparatus.
Some or all of the method steps may be executed by (or using) a hardware apparatus,
like for example, a microprocessor, a programmable computer or an electronic circuit.
In some embodiments, some one or more of the most important method steps may be
executed by such an apparatus.
The inventive encoded audio signal can be stored on a digital storage medium or can be
transmitted on a transmission medium such as a wireless transmission medium or a
wired transmission medium such as the Internet.
Depending on certain implementation requirements, embodiments of the invention can
be implemented in hardware or in software. The implementation can be performed
using a digital storage medium, for example a floppy disk, a DVD, a Blue-Ray, a CD, a
ROM, a PROM, an EPROM, an EEPROM or a FLASH memory, having electronically
readable control signals stored thereon, which cooperate (or are capable of cooperating)
with a programmable computer system such that the respective method is performed.
Therefore, the digital storage medium may be computer readable.
Some embodiments according to the invention comprise a data carrier having
electronically readable control signals, which are capable of cooperating with a
programmable computer system, such that one of the methods described herein is
performed.
Generally, embodiments of the present invention can be implemented as a computer
program product with a program code, the program code being operative for performing
one of the methods when the computer program product runs on a computer. The
program code may for example be stored on a machine readable carrier.
Other embodiments comprise the computer program for performing one of the methods
described herein, stored on a machine readable carrier.
In other words, an embodiment of the inventive method is, therefore, a computer
program having a program code for performing one of the methods described herein,
when the computer program runs on a computer.
A further embodiment of the inventive methods is, therefore, a data carrier (or a digital
storage medium, or a computer-readable medium) comprising, recorded thereon, the
computer program for performing one of the methods described herein. The data carrier,
the digital storage medium or the recorded medium are typically tangible and/or non-
transitionary.
A further embodiment of the inventive method is, therefore, a data stream or a sequence
of signals representing the computer program for performing one of the methods
described herein. The data stream or the sequence of signals may for example be
configured to be transferred via a data communication connection, for example via the
Internet.
A further embodiment comprises a processing means, for example a computer, or a
programmable logic device, configured to or adapted to perform one of the methods
described herein.
A further embodiment comprises a computer having installed thereon the computer
program for performing one of the methods described herein.
In some embodiments, a programmable logic device (for example a field programmable
gate array) may be used to perform some or all of the functionalities of the methods
described herein. In some embodiments, a field programmable gate array may cooperate
with a microprocessor in order to perform one of the methods described herein.
Generally, the methods are preferably performed by any hardware apparatus.
The above described embodiments are merely illustrative for the principles of the
present invention. It is understood that modifications and variations of the arrangements
and the details described herein will be apparent to others skilled in the art. It is the
intent, therefore, to be limited only by the scope of the impending patent claims and not
by the specific details presented by way of description and explanation of the
embodiments herein.
References
[1] C. Faller and F. Baumgarte, "Binaural Cue Coding - Part II: Schemes and
applications", IEEE Trans, on Speech and Audio Proc, vol. 11, no. 6, Nov.
2003.
[2] C. Faller, "Parametric Joint-Coding of Audio Sources", 120th AES Convention,
Paris, 2006, Preprint 6752.
[3] J. Herre, S. Disch, J. Hilpert, O. Hellmuth: "From SAC To SAOC - Recent
Developments in Parametric Coding of Spatial Audio", 22nd Regional UK AES
Conference, Cambridge, UK, April 2007.
[4] J. Engdegard, B. Resch, C. Falch, O. Hellmuth, J. Hilpert, A. HSlzer, L.
Terentiev, J. Breebaart, J. Koppens, E. Schuijers and W. Oomen: "Spatial Audio
Object Coding (SAOC) - The Upcoming MPEG Standard on Parametric Object
Based Audio Coding", 124th AES Convention, Amsterdam 2008, Preprint 7377.
[5] ISO/IEC, "MPEG audio technologies - Part 2: Spatial Audio Object Coding
(SAOC)," ISO/IEC JTC1/SC29/WG11 (MPEG) FCD 23003-2.
[6] US patent application 61/173,456, METHODS, APPARATUS, AND
COMPUTER PROGRAMS FOR DISTORTION AVOIDING AUDIO SIGNAL
PROCESSING
[7] EBU Technical recommendation: "MUSHRA-EBU Method for Subjective
Listening Tests of Intermediate Audio Quality", Doc. B/AIM022, October 1999.
[8] ISO/IEC JTC1/SC29/WG11 (MPEG), Document N10843, "Study on ISO/IEC
23003-2:200x Spatial Audio Object Coding (SAOC)", 89th MPEG Meeting,
London, UK, July 2009
We Claim:
1. An audio processing apparatus (100; 200) for providing an upmix signal
representation (130; 230) on the basis of a downmix signal representation (110;
210)and an object-related parametric information, which are included in a
bitstream representation (300) of an audio content, and in dependence on a user-
specified rendering matrix (144, Mren) which defines a desired contribution of a
plurality of audio objects to one, two or more output audio channels, the apparatus
comprising;
a distortion limiter (140; 240) configured to obtain a modified rendering matrix
(142; Mren]im) using a linear combination of a user-specified rendering matrix
(Mren) and a distortion-free target rendering matrix (Mren ar) in dependence on a
linear combination parameter (146; gDCU); and
a signal processor (148; 248) configured to obtain the upmix signal representation
on the basis of the downmix signal representation and the object-related
parametric information using the modified rendering matrix;
wherein the apparatus is configured to evaluate a bitstream element (306;
bsDcuParameter) representing the linear combination parameter (146; gDCU) in
order to obtain the linear combination parameter,
2. The apparatus (100; 200) according to claim 1, wherein the distortion limiter is
configured to obtain the target rendering matrix (Mren,lar) such that the target
rendering matrix is a distortion-free target rendering matrix.
3. The apparatus (100; 200) according to claim 1 or claim 2, wherein the distortion
limiter is configured to obtain the modified rendering matrix according to:
wherein gocu designates the linear combination parameter, a value of which is in
an interval [0,1];
wherein designates the user-specified rendering matrix; and
FH101106PCT-2012001818
wherein designates the target rendering matrix.
4. The apparatus (100; 200) according to one of claims 1 to 3, wherein the distortion
limiter is configured to obtain the target rendering matrix (M renMr) such that the
target rendering matrix is a downmix - similar target rendering matrix.
5. The apparatus (100; 200) according to one of claims 1 to 4, wherein the distortion
limiter is configured to scale an extended downmix matrix (D'DS) using an energy
normalization scalar to obtain the target rendering matrix (JM.ren,iarwherein the extended downmix matrix is an extended version of a downmix
matrix, one or more rows of which downmix matrix describe contributions of a
plurality of audio object signals to one or more channels of the downmix signal
representation, extended by rows of zero elements, such that a number of rows of
the extended downmix matrix is identical to a rendering constellation described
by the user-specified rendering matrix (Mren).
6. The apparatus (100; 200) according to one of claims 1 to 3, wherein the distortion
limiter is configured to obtain the target rendering matrix (M ren,tar), such that the
target rendering matrix is a best-effort target rendering matrix.
7. The apparatus (100; 200) according to one of claims 1 to 3 or 6, wherein the
distortion limiter is configured to obtain the target rendering matrix (M ren,tar)
such that the target rendering matrix depends on a downmix matrix (D) and the
user specified rendering matrix (Mren).
8. The apparatus (100; 200) according to one of claims 1 to 3, 6 or 7, wherein the
distortion limiter is configured to compute a matrix (NM) comprising channel
individual energy normalization values for a plurality of output audio channels of
the apparatus for providing an upmix signal representation, such that an energy
normalization value for a given output audio channel of the apparatus describes, at
least approximately, a ratio between a sum of energy rendering values associated
with the given output audio channel in the user-specified rendering matrix for a
plurality of audio objects and a sum of energy downmix values for the plurality of
audio objects; and
wherein the distortion limiter is configured to scale a set of downmix values using
channel-individual energy normalization value, to obtain a set of rendering values
of the target rendering matrix (M rertter) associated with the given output channel.
9. The apparatus (100; 200) according to one of claims 1 to 3 and 6 to 8, wherein the
distortion limiter is configured to compute a matrix (N'l,m) comprising channel-
individual energy normalization values for a plurality of output audio channels
according to:
for the case of a 1-channel downmix signal representation and a 2-channel output
signal of the apparatus; or
according to:
for the case of a 1-channel downmix signal representation and a binaural-rendered
output signal of the apparatus; or
according to:
for the case of a 1-channel downmix signal representation and a N MPS -channel
output signal of the apparatus;
wherein designates rendering coefficients of the user-specified rendering
matrix describing a desired contribution of an audio object having object
index j to a first output audio channel of the apparatus;
wherein designates rendering coefficients of the user-specified rendering
matrix describing a desired contribution of an audio object having object
index j to a second output audio channel of the apparatus;
wherein and designate the rendering coefficients of the user-specified
rendering matrix describing a desired contribution of an audio object
having object index j to a first and second output audio channel of the apparatus,
and taking parametric HRTF information into consideration.
wherein designates a downmix coefficient describing a contribution of an
audio object having an object index j to the downmix signal representation; and
wherein designates an additive constant to avoid division by zero; and
wherein the distortion limiter is configured to compute the target rendering matrix
according to:
wherein D1 designates a downmix matrix comprising the downmix coefficient dj.
10. The apparatus (100; 200) according to one of claims 1 to 3 or 6 to 7, wherein the
distortion limiter is configured to compute a matrix describing a channel-
individual energy normalization for a plurality of output audio channels of the
apparatus in dependence on the user-specified rendering matrix and a
downmix matrix D; and
wherein the distortion limiter is configured to apply the matrix describing the
channel-individual energy normalization to obtain a set of rendering coefficients
of the target rendering matrix associated with a given output audio
channel of the apparatus as a linear combination of sets of downmix values
associated with different channels of the downmix signal representation.
11. The apparatus (100; 200) according to one of claims 1 to 3 or 6 to 7, or 10,
wherein the distortion limiter is configured to compute a matrix describing
the channel-individual energy normalization for a plurality of output audio
channels according to:
for the case of a 2-channel downmix signal representation and a multi-channel
output audio signal of the apparatus;
wherein designates the user-specified rendering matrix describing user-
specified, desired contributions of a plurality of audio object signals to the multi-
channel output audio signal of the apparatus;
wherein D' designates a downmix matrix describing contributions of a plurality
of audio object signals to the downmix signal representation;
wherein
wherein the distortion limiter is configured to compute the target rendering matrix
according to
12. The apparatus (100; 200) according to claims 1 to 3 or 6 to 7, or 10, wherein the
distortion limiter is configured to compute a matrix according to
for the case of a 2-channel downmix signal representation and a 1-channel output
audio signal of the apparatus, or
according to
for the case of a 2-channel downmix signal representation and a binaurally-
rendered output audio signal of the apparatus;
wherein designates the user-specified rendering matrix describing user-
specified desired contributions of a plurality of audio object signals to the output
signal of the apparatus;
wherein D' designates a downmix matrix describing contributions of a plurality of
audio object signals to the downmix signal representation;
wherein designates a binaural rendering matrix which is based on the user-
specified rendering matrix and parameters of a head-related transfer function.
13. The apparatus (100; 200) according to one of claims 1 to 3 or 6 to 7, wherein the
distortion limiter is configured to compute an energy normalization scalar
according to
wherein designates a rendering coefficient of the user-specified rendering
matrix describing a desired contribution of an audio object having object
index j to an output audio signal of the apparatus;
wherein designates a downmix coefficient describing a contribution of an
audio object having object index j to the downmix signal representation; and
wherein e designates an additive constant to avoid division by zero.
14. The apparatus (100; 200) according to one of claims 1 to 13, wherein the
apparatus is configured to read an index value (idx) representing the linear
combination parameter (gDCU) from the bitstream representation of the audio
content and to map the index value onto the linear combination parameter (gDCU)
using a parameter quantization table.
15. The apparatus (100; 200) according to claim 14, wherein the quantization table
describes a non-uniform quantization, wherein smaller values of the linear
combination parameter (gDCU), which describe a stronger contribution of the
user-specified rendering matrix (Mren) onto the modified rendering matrix
(M ren,lim), are quantized with higher resolution.
16. The apparatus (100; 200) according to one of claims 1 to 15, wherein the
apparatus is configured to evaluate a bitstream element (bsDcuMode) describing a
distortion limitation mode, and wherein the distortion limiter is configured to
selectively obtain the target rendering matrix such that the target rendering matrix
is a downmix-similar target rendering matrix, or such that the target rendering
matrix is a best-effort target rendering matrix.
17. An apparatus (150) for providing a bitstream (170) representing a multi-channel
audio signal, the apparatus comprising:
a downmixer (180) configured to provide a downmix signal (182) on the basis of
a plurality of audio object signals (160a-160N);
a side information provider (184) configured to provide an object-related
parametric side information (186) describing characteristics of the audio object
signals (160a-160N) and downmix parameters, and a linear combination
parameter (188) describing desired contributions of a user-specified rendering
matrix (M ren) and of a target rendering matrix (M ren,tar) to a modified rendering
matrix (Mren,lim) to be used by an apparatus (100; 200) for providing an upmix
signal representation on the basis of the bitstream; and
a bitstream formatter (190) configured to provide a bitstream (170) comprising a
representation of the downmix signal, of the object-related parametric side
information and of the linear combination parameter;
wherein the user-specified rendering matrix (144, Mren) defines a desired
contribution of a plurality of audio objects to one, two or more output audio
channels.
18. An audio processing method for providing an upmix signal representation on the
basis of a downmix signal representation and an object-related parametric
information, which are included in a bitstream representation of an audio content,
and in a dependence on a user-specified rendering matrix which defines a desired
contribution of a plurality of audio objects to one, two or more output audio
channels, the method comprising:
evaluating a bitstream element representing a linear combination parameter, in
order to obtain the linear combination parameter;
obtaining a modified rendering matrix using a linear combination of a user-
specified rendering matrix and a distortion-free target rendering matrix in
dependence on the linear combination parameter; and
obtaining the upmix signal representation on the basis of the downmix signal
representation and the object-related parametric information using the modified
rendering matrix.
19. A method for providing a bitstream representing a multi-channel audio signal, the
method comprising:
providing a downmix signal on the basis of a plurality of audio object signals;
providing an object-related parametric side information describing characteristics
of the audio object signals and downmix parameters, and a linear combination
parameter describing desired contributions of a user-specified rendering matrix
and of a target rendering matrix to a modified rendering matrix; and
providing a bitstream comprising a representation of the downmix signal, of the
object-related parametric side information and the linear combination parameter;
wherein the user-specified rendering matrix defines a desired contribution of a
plurality of audio objects to one, two or more output audio channels.
20. A computer program for performing a method according to claim 18 or 19 when
the computer program runs on a computer.
21. A bitstream (300) representing a multi-channel audio signal, the bitstream
comprising:
a representation (302) of a downmix signal combining audio signals of a plurality
of audio objects:
an object-related parametric information (304) describing characteristics of the
audio objects; and
a linear combination parameter (306) describing desired contributions of a user-
specified rendering matrix and of a target rendering matrix to a modified
rendering matrix.
ABSTRACT
An apparatus for providing an upmix signal representation on the basis of a downmix
signal representation and an object-related parametric information, which are included
in a bitstream representation of an audio content, in independence on a user-specified
rendering matrix, the apparatus comprises a distortion limiter configured to obtain a
modified rendering matrix using a linear combination of a user-specified rendering
matrix in a target rendering matrix in dependence on a linear combination parameter.
The apparatus also comprises a signal processor configured to obtain the upmix signal
representation on the basis of the downmix signal representation and the object-related
parametric information using the modified rendering matrix. The apparatus is also
configured to evaluate a bitstream element representing the linear combination
parameter in order to obtain the linear combination parameter.
| # | Name | Date |
|---|---|---|
| 1 | 1166-KOLNP-2012-(16-05-2012)-SPECIFICATION.pdf | 2012-05-16 |
| 1 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [25-09-2023(online)].pdf | 2023-09-25 |
| 2 | 1166-KOLNP-2012-(16-05-2012)-PCT SEARCH REPORT & OTHERS.pdf | 2012-05-16 |
| 2 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [08-09-2023(online)].pdf | 2023-09-08 |
| 3 | 1166-KOLNP-2012-PROOF OF ALTERATION [24-05-2023(online)].pdf | 2023-05-24 |
| 3 | 1166-KOLNP-2012-(16-05-2012)-OTHERS.pdf | 2012-05-16 |
| 4 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [27-09-2022(online)].pdf | 2022-09-27 |
| 4 | 1166-KOLNP-2012-(16-05-2012)-INTERNATIONAL PUBLICATION.pdf | 2012-05-16 |
| 5 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [06-09-2022(online)].pdf | 2022-09-06 |
| 5 | 1166-KOLNP-2012-(16-05-2012)-GPA.pdf | 2012-05-16 |
| 6 | 1166-KOLNP-2012-IntimationOfGrant14-12-2020.pdf | 2020-12-14 |
| 6 | 1166-KOLNP-2012-(16-05-2012)-FORM-5.pdf | 2012-05-16 |
| 7 | 1166-KOLNP-2012-PatentCertificate14-12-2020.pdf | 2020-12-14 |
| 7 | 1166-KOLNP-2012-(16-05-2012)-FORM-3.pdf | 2012-05-16 |
| 8 | 1166-KOLNP-2012-Information under section 8(2) [03-11-2020(online)].pdf | 2020-11-03 |
| 8 | 1166-KOLNP-2012-(16-05-2012)-FORM-2.pdf | 2012-05-16 |
| 9 | 1166-KOLNP-2012-(16-05-2012)-FORM-1.pdf | 2012-05-16 |
| 9 | 1166-KOLNP-2012-FORM 3 [04-06-2020(online)].pdf | 2020-06-04 |
| 10 | 1166-KOLNP-2012-(16-05-2012)-DRAWINGS.pdf | 2012-05-16 |
| 10 | 1166-KOLNP-2012-Information under section 8(2) [22-02-2020(online)].pdf | 2020-02-22 |
| 11 | 1166-KOLNP-2012-(16-05-2012)-DESCRIPTION (COMPLETE).pdf | 2012-05-16 |
| 11 | 1166-KOLNP-2012-Information under section 8(2) [18-02-2020(online)].pdf | 2020-02-18 |
| 12 | 1166-KOLNP-2012-(16-05-2012)-CORRESPONDENCE.pdf | 2012-05-16 |
| 12 | 1166-KOLNP-2012-Information under section 8(2) (MANDATORY) [10-05-2019(online)].pdf | 2019-05-10 |
| 13 | 1166-KOLNP-2012-(16-05-2012)-CLAIMS.pdf | 2012-05-16 |
| 13 | 1166-KOLNP-2012-Information under section 8(2) (MANDATORY) [07-01-2019(online)].pdf | 2019-01-07 |
| 14 | 1166-KOLNP-2012-(16-05-2012)-AMENDED CLAIMS.pdf | 2012-05-16 |
| 14 | 1166-KOLNP-2012-ABSTRACT [07-06-2018(online)].pdf | 2018-06-07 |
| 15 | 1166-KOLNP-2012-(16-05-2012)-ABSTRACT.pdf | 2012-05-16 |
| 15 | 1166-KOLNP-2012-CLAIMS [07-06-2018(online)].pdf | 2018-06-07 |
| 16 | 1166-KOLNP-2012-(03-08-2012)-PA.pdf | 2012-08-03 |
| 16 | 1166-KOLNP-2012-CORRESPONDENCE [07-06-2018(online)].pdf | 2018-06-07 |
| 17 | 1166-KOLNP-2012-DRAWING [07-06-2018(online)].pdf | 2018-06-07 |
| 17 | 1166-KOLNP-2012-(03-08-2012)-CORRESPONDENCE.pdf | 2012-08-03 |
| 18 | 1166-KOLNP-2012-(08-10-2012)-CORRESPONDENCE.pdf | 2012-10-08 |
| 18 | 1166-KOLNP-2012-FER_SER_REPLY [07-06-2018(online)].pdf | 2018-06-07 |
| 19 | 1166-KOLNP-2012-(08-10-2012)-ASSIGNMENT.pdf | 2012-10-08 |
| 19 | 1166-KOLNP-2012-PETITION UNDER RULE 137 [07-06-2018(online)].pdf | 2018-06-07 |
| 20 | 1166-KOLNP-2012-(14-11-2012)-CORRESPONDENCE.pdf | 2012-11-14 |
| 20 | 1166-KOLNP-2012-FER.pdf | 2017-12-07 |
| 21 | 1166-KOLNP-2012-(14-11-2012)-ANNEXURE TO FORM 3.pdf | 2012-11-14 |
| 21 | 1166-KOLNP-2012-Information under section 8(2) (MANDATORY) [06-09-2017(online)].pdf | 2017-09-06 |
| 22 | 1166-KOLNP-2012-(29-01-2013)-OTHERS.pdf | 2013-01-29 |
| 22 | Other Patent Document [20-03-2017(online)].pdf | 2017-03-20 |
| 23 | 1166-KOLNP-2012-(29-01-2013)-CORRESPONDENCE.pdf | 2013-01-29 |
| 23 | Other Patent Document [16-09-2016(online)].pdf | 2016-09-16 |
| 24 | Other Patent Document [16-09-2016(online)].pdf | 2016-09-16 |
| 24 | 1166-KOLNP-2012-(29-01-2013)-CORRESPONDENCE.pdf | 2013-01-29 |
| 25 | 1166-KOLNP-2012-(29-01-2013)-OTHERS.pdf | 2013-01-29 |
| 25 | Other Patent Document [20-03-2017(online)].pdf | 2017-03-20 |
| 26 | 1166-KOLNP-2012-(14-11-2012)-ANNEXURE TO FORM 3.pdf | 2012-11-14 |
| 26 | 1166-KOLNP-2012-Information under section 8(2) (MANDATORY) [06-09-2017(online)].pdf | 2017-09-06 |
| 27 | 1166-KOLNP-2012-(14-11-2012)-CORRESPONDENCE.pdf | 2012-11-14 |
| 27 | 1166-KOLNP-2012-FER.pdf | 2017-12-07 |
| 28 | 1166-KOLNP-2012-(08-10-2012)-ASSIGNMENT.pdf | 2012-10-08 |
| 28 | 1166-KOLNP-2012-PETITION UNDER RULE 137 [07-06-2018(online)].pdf | 2018-06-07 |
| 29 | 1166-KOLNP-2012-(08-10-2012)-CORRESPONDENCE.pdf | 2012-10-08 |
| 29 | 1166-KOLNP-2012-FER_SER_REPLY [07-06-2018(online)].pdf | 2018-06-07 |
| 30 | 1166-KOLNP-2012-(03-08-2012)-CORRESPONDENCE.pdf | 2012-08-03 |
| 30 | 1166-KOLNP-2012-DRAWING [07-06-2018(online)].pdf | 2018-06-07 |
| 31 | 1166-KOLNP-2012-(03-08-2012)-PA.pdf | 2012-08-03 |
| 31 | 1166-KOLNP-2012-CORRESPONDENCE [07-06-2018(online)].pdf | 2018-06-07 |
| 32 | 1166-KOLNP-2012-(16-05-2012)-ABSTRACT.pdf | 2012-05-16 |
| 32 | 1166-KOLNP-2012-CLAIMS [07-06-2018(online)].pdf | 2018-06-07 |
| 33 | 1166-KOLNP-2012-(16-05-2012)-AMENDED CLAIMS.pdf | 2012-05-16 |
| 33 | 1166-KOLNP-2012-ABSTRACT [07-06-2018(online)].pdf | 2018-06-07 |
| 34 | 1166-KOLNP-2012-(16-05-2012)-CLAIMS.pdf | 2012-05-16 |
| 34 | 1166-KOLNP-2012-Information under section 8(2) (MANDATORY) [07-01-2019(online)].pdf | 2019-01-07 |
| 35 | 1166-KOLNP-2012-(16-05-2012)-CORRESPONDENCE.pdf | 2012-05-16 |
| 35 | 1166-KOLNP-2012-Information under section 8(2) (MANDATORY) [10-05-2019(online)].pdf | 2019-05-10 |
| 36 | 1166-KOLNP-2012-Information under section 8(2) [18-02-2020(online)].pdf | 2020-02-18 |
| 36 | 1166-KOLNP-2012-(16-05-2012)-DESCRIPTION (COMPLETE).pdf | 2012-05-16 |
| 37 | 1166-KOLNP-2012-(16-05-2012)-DRAWINGS.pdf | 2012-05-16 |
| 37 | 1166-KOLNP-2012-Information under section 8(2) [22-02-2020(online)].pdf | 2020-02-22 |
| 38 | 1166-KOLNP-2012-(16-05-2012)-FORM-1.pdf | 2012-05-16 |
| 38 | 1166-KOLNP-2012-FORM 3 [04-06-2020(online)].pdf | 2020-06-04 |
| 39 | 1166-KOLNP-2012-(16-05-2012)-FORM-2.pdf | 2012-05-16 |
| 39 | 1166-KOLNP-2012-Information under section 8(2) [03-11-2020(online)].pdf | 2020-11-03 |
| 40 | 1166-KOLNP-2012-(16-05-2012)-FORM-3.pdf | 2012-05-16 |
| 40 | 1166-KOLNP-2012-PatentCertificate14-12-2020.pdf | 2020-12-14 |
| 41 | 1166-KOLNP-2012-(16-05-2012)-FORM-5.pdf | 2012-05-16 |
| 41 | 1166-KOLNP-2012-IntimationOfGrant14-12-2020.pdf | 2020-12-14 |
| 42 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [06-09-2022(online)].pdf | 2022-09-06 |
| 42 | 1166-KOLNP-2012-(16-05-2012)-GPA.pdf | 2012-05-16 |
| 43 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [27-09-2022(online)].pdf | 2022-09-27 |
| 43 | 1166-KOLNP-2012-(16-05-2012)-INTERNATIONAL PUBLICATION.pdf | 2012-05-16 |
| 44 | 1166-KOLNP-2012-PROOF OF ALTERATION [24-05-2023(online)].pdf | 2023-05-24 |
| 44 | 1166-KOLNP-2012-(16-05-2012)-OTHERS.pdf | 2012-05-16 |
| 45 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [08-09-2023(online)].pdf | 2023-09-08 |
| 45 | 1166-KOLNP-2012-(16-05-2012)-PCT SEARCH REPORT & OTHERS.pdf | 2012-05-16 |
| 46 | 1166-KOLNP-2012-RELEVANT DOCUMENTS [25-09-2023(online)].pdf | 2023-09-25 |
| 46 | 1166-KOLNP-2012-(16-05-2012)-SPECIFICATION.pdf | 2012-05-16 |
| 1 | Search(15)_22-02-2017.pdf |