Title of Invention

"ÄUDIO DECODING METHOD AND AUDIO DECODING APPARATUS"

Abstract Provided are an audio encoding method and apparatus and an audio decoding method and apparatus in which audio signals can be encoded or decoded so that sound images can be localized at any desired position for each object audio signal. The audio decoding method includes extracting a downmix signal and object-based side information from an input audio signal; generating rendering information based on input control data; and generating spatial information based on the rendering information and the object-based side information.
Full Text Description
METHODS AND APPARATUSES FOR ENCODING AND
DECODING OBJECT-BASED AUDIO SIGNALS
Technical Field
The audio signal decoding method includes extracting a downmix signal and object-
based side information from an audio signal; generating a modified downmix signal
based on the downmix signal and extracted information which is extracted from the
object-based side information; generating channel-based side information based on the
object-based side information and control data for rendering the downmix signal; and
generating a multi-channel audio signal based on the modified downmix signal and the
channel-based side information.
Background Art
In general, in multi-channel audio encoding and decoding techniques, a number of
channel signals of a multi-channel signal are downmixed into fewer channel signals,
side information regarding the original channel signals is transmitted, and a multichannel
signal having as many channels as the original multi-channel signal is
restored.
Object-based audio encoding and decoding techniques are basically similar to multichannel
audio encoding and decoding techniques in terms of downmixing several
sound sources into fewer sound source signals and transmitting side information
regarding the original sound sources. However, in object-based audio encoding and
decoding techniques, object signals, which are basic elements (e.g., the sound of a
musical instrument or a human voice) of a channel signal, are treated the same as
channel signals in multi-channel audio encoding and decoding techniques and can thus
be coded.
In other words, in object-based audio encoding and decoding techniques, each object
signal is deemed the entity to be coded. In this regard, object-based audio encoding and
decoding techniques are different from multi-channel audio encoding and decoding
techniques in which a multi-channel audio coding operation is performed simply based
on inter-channel information regardless of the number of elements of a channel signal
to be coded.
Disclosure of Invention
Technical Problem
The present invention provides an audio encoding method and apparatus and an
audio decoding method and apparatus in which audio signals can be encoded or

decoded so that sound images can be localized at any desired position for each object
audio signal.
Technical Solution
According to an aspect of the present invention, there is provided an audio decoding
method including extracting a downmix signal and object-based side information from
an input audio signal; generating rendering information based on input control data;
and generating spatial information based on the rendering information and the object-
based side information.
According to another aspect of the present invention, there is provided an audio
decoding apparatus including a demultiplexer which extracts a downmix signal and
object-based side information from an input audio signal; a renderer which generates
rendering information based on input control data; and a transcoder which generates
spatial information based on the rendering information and the object-based side information.
According to an aspect of the present invention, there is provided a computer-
readable recording medium having recorded thereon a computer program for executing
an audio decoding method, the audio decoding method including extracting a downmix
signal and object-based side information from an input audio signal; generating
rendering information based on input control data; and generating spatial information
based on the rendering information and the object-based side information
Advantageous Effects
Provided are an audio encoding method and apparatus and an audio decoding method
and apparatus in which audio signals can be encoded or decoded so that sound images
can be localized at any desired position for each object audio signal.
Brief Description of the Drawings
The present invention will become more fully understood from the detailed de-
scription given hereinbelow and the accompanying drawings, which are given by il-
lustration only, and thus are not limitative of the present invention, and wherein:
FIG. 1 is a block diagram of a typical object-based audio encoding/decoding system;
FIG. 2 is a block diagram of an audio decoding apparatus according to a first
embodiment of the present invention;
FIG. 3 is a block diagram of an audio decoding apparatus according to a second
embodiment of the present invention;
FIG. 4 is a graph for explaining the influence of an amplitude difference and a time
difference, which are independent from each other, on the localization of sound
images;

FIG. 5 is a graph of functions regarding the correspondence between amplitude
differences and time differences which are required to localize sound images at a predetermined
position;
FIG. 6 illustrates the format of control data including harmonic information;
FIG. 7 is a block diagram of an audio decoding apparatus according to a third
embodiment of the present invention;
FIG. 8 is a block diagram of an artistic downmix gains (ADG) module that can be
used in the audio decoding apparatus illustrated in FIG. 7;
FIG. 9 is a block diagram of an audio decoding apparatus according to a fourth
embodiment of the present invention;
FIG. 10 is a block diagram of an audio decoding apparatus according to a fifth
embodiment of the present invention;
FIG. 11 is a block diagram of an audio decoding apparatus according to a sixth
embodiment of the present invention;
FIG. 12 is a block diagram of an audio decoding apparatus according to a seventh
embodiment of the present invention;
FIG. 13 is a block diagram of an audio decoding apparatus according to an eighth
embodiment of the present invention;
FIG. 14 is a diagram for explaining the application of three-dimensional (3D) information
to a frame by the audio decoding apparatus illustrated in FIG. 13;
FIG. 15 is a block diagram of an audio decoding apparatus according to a ninth
embodiment of the present invention;
FIG. 16 is a block diagram of an audio decoding apparatus according to a tenth
embodiment of the present invention;
FIGS. 17 through 19 are diagrams for explaining an audio decoding method
according to an embodiment of the present invention; and
FIG. 20 is a block diagram of an audio encoding apparatus according to an
embodiment of the present invention.
Best Mode for Carrying Out the Invention
The present invention will hereinafter be described in detail with reference to the accompanying
drawings in which exemplary embodiments of the invention are shown.
An audio encoding method and apparatus and an audio decoding method and
apparatus according to the present invention may be applied to object- based audio
processing operations, but the present invention is not restricted to this. In other words,
the audio encoding method and apparatus and the audio decoding method and
apparatus may be applied to various signal processing operations other than object-
based audio processing operations.

FIG. 1 is a block diagram of a typical object-based audio encoding/decoding system.
In general, audio signals input to an object-based audio encoding apparatus do not
correspond to channels of a multi-channel signal but are independent object signals. In
this regard, an object-based audio encoding apparatus is differentiated from a multichannel
audio encoding apparatus to which channel signals of a multi-channel signal
are input.
For example, channel signals such as a front left channel signal and a front right
channel signal of a 5.1-channel signal may be input to a multi-channel audio signal,
whereas object audio signals such as a human voice or the sound of a musical
instrument (e.g., the sound of a violin or a piano) which are smaller entities than
channel signals may be input to an object-based audio encoding apparatus.
Referring to FIG. 1, the object-based audio encoding/decoding system includes an
object-based audio encoding apparatus and an object-based audio decoding apparatus.
The object-based audio encoding apparatus includes an object encoder 100, and the
object-based audio decoding apparatus includes an object decoder 111 and a renderer
113.
The object encoder 100 receives N object audio signals, and generates an object-
based downmix signal with one or more channels and side information including a
number of pieces of information extracted from the N object audio signals such as
energy difference, phase difference, and correlation value. The side information and
the object-based downmix signal are incorporated into a single bitstream, and the
bitstream is transmitted to the object-based decoding apparatus.
The side information may include a flag indicating whether to perform channel-based
audio coding or object-based audio coding, and thus, it may be determined whether to
perform channel-based audio coding or object-based audio coding based on the flag of
the side information. The side information may also include envelope information,
grouping information, silent period information, and delay information regarding
object signals. The side information may also include object level differences
informaion, inter-object cross correlation information, downmix gain information,
downmix channel level difference information, and absolute object energy in-
foramtion.
The object decoder 111 receives the object-based downmix signal and the side information
from the object-based audio encoding apparatus, and restores object signals
having similar properties to those of the N object audio signals based on the object-
based downmix signal and the side information. The object signals generated by the
object decoder 111 have not yet been allocated to any position in a multi-channel
space. Thus, the renderer 113 allocates each of the object signals generated by the
object decoder 111 to a predetermined position in a multi-channel space and

determines the levels of the object signals so that the object signals can be reproduced
from respective corresponding positions designated by the renderer 113 with respective
corresponding levels determined by the renderer 113. Control information regarding
each of the object signals generated by the object decoder 111 may vary over time, and
thus, the spatial positions and the levels of the object signals generated by the object
decoder 111 may vary according to the control information.
FIG. 2 is a block diagram of an audio decoding apparatus 120 according to a first
embodiment of the present invention. Referring to FIG. 2, the audio decoding
apparatus 120 includes an object decoder 121, a renderer 123, and a parameter
converter 125. The audio decoding apparatus 120 may also include a demultiplexer
(not shown) which extracts a downmix signal and side information from a bitstream
input thereto, and this will apply to all audio decoding apparatuses according to other
embodiments of the present invention.
The object decoder 121 generates a number of object signals based on a downmix
signal and modified side information provided by the parameter converter 125. The
renderer 123 allocates each of the object signals generated by the object decoder 121 to
a predetermined position in a multi-channel space and determines the levels of the
object signals generated by the object decoder 121 according to control information.
The parameter converter 125 generates the modified side information by combining
the side information and the control information. Then, the parameter converter 125
transmits the modified side information to the object decoder 121.
The object decoder 121 may be able to perform adaptive decoding by analyzing the
control information in the modified side information.
For example, if the control information indicates that a first object signal and a
second object signal are allocated to the same position in a multi-channel space and
have the same level, a typical audio decoding apparatus may decode the first and
second object signals separately, and then arrange them in a multi-channel space
through a mixing/rendering operation.
On the other hand, the object decoder 121 of the audio decoding apparatus 120 learns
from the control information in the modified side information that the first and second
object signals are allocated to the same position in a multi-channel space and have the
same level as if they were a single sound source. Accordingly, the object decoder 121
decodes the first and second object signals by treating them as a single sound source
without decoding them separately. As a result, the complexity of decoding decreases.
In addition, due to a decrease in the number of sound sources that need to be
processed, the complexity of mixing/rendering also decreases.
The audio decoding apparatus 120 may be effectively used in the situation when the
number of object signals is greater than the number of output channels because a

plurality of object signals are highly likely to be allocated to the same spatial position.
Alternatively, the audio decoding apparatus 120 may be used in the situation when
the first object signal and the second object signal are allocated to the same position in
a multi-channel space but have different levels. In this case, the audio decoding
apparatus 120 decode the first and second object signals by treating the first and
second object signals as a single, instead of decoding the first and second object
signals separately and transmitting the decoded first and second object signals to the
Tenderer 123. More specifically, the object decoder 121 may obtain information
regarding the difference between the levels of the first and second object signals from
the control information in the modified side information, and decode the first and
second object signals based on the obtained information. As a result, even if the first
and second object signals have different levels, the first and second object signals can
be decoded as if they were a single sound source.
Still alternatively, the object decoder 121 may adjust the levels of the object signals
generated by the object decoder 121 according to the control information. Then, the
object decoder 121 may decode the object signals whose levels are adjusted. Ac-
cordingly, the renderer 123 does not need to adjust the levels of the decoded object
signals provided by the object decoder 121 but simply arranges the decoded object
signals provided by the object decoder 121 in a multi-channel space. In short, since the
object decoder 121 adjusts the levels of the object signals generated by the object
decoder 121 according to the control information, the Tenderer 123 can readily arrange
the object signals generated by the object decoder 121 in a multi-channel space without
the need to additionally adjust the levels of the object signals generated by the object
decoder 121. Therefore, it is possible to reduce the complexity of mixing/rendering.
According to the embodiment of FIG. 2, the object decoder of the audio decoding
apparatus 120 can adaptively perform a decoding operation through the analysis of the
control information, thereby reducing the complexity of decoding and the complexity
of mixing/rendering. A combination of the above-described methods performed by the
audio decoding apparatus 120 may be used.
FIG. 3 is a block diagram of an audio decoding apparatus 130 according to a second
embodiment of the present invention. Referring to FIG. 3, the audio decoding
apparatus 130 includes an object decoder 131 and a Tenderer 133. The audio decoding
apparatus 130 is characterized by providing side information not only to the object
decoder 131 but also to the Tenderer 133.
The audio decoding apparatus 130 may effectively perform a decoding operation
even when there is an object signal corresponding to a silent period. For example,
second through fourth object signals may correspond to a music play period during
which a musical instrument is played, and a first object signal may correspond to a

silent period during which an accompaniment is played. In this case, information
indicating which of a plurality of object signals corresponds to a silent period may be
included in side information, and the side information may be provided to the renderer
133 as well as to the object decoder 131.
The object decoder 131 may minimize the complexity of decoding by not decoding
an object signal corresponding to a silent period. The object decoder 131 sets an object
signal corresponding to a value of 0 and transmits the level of the object signal to the
renderer 133. In general, object signals having a value of 0 are treated the same as
object signals having a value, other than 0, and are thus subjected to a mixing/
rendering operation.
On the other hand, the audio decoding apparatus 130 transmits side information
including information indicating which of a plurality of object signals corresponds to a
silent period to the renderer 133 and can thus prevent an object signal corresponding to
a silent period from being subjected to a mixing/rendering operation performed by the
renderer 133. Therefore, the audio decoding apparatus 130 can prevent an unnecessary
increase in the complexity of mixing/rendering.
The renderer 133 may use mixing parameter information which is included in control
information to localize a sound image of each object signal at a stereo scene. The
mixing parameter information may include amplitude information only or both
amplitude information and time information. The mixing parameter information affects
not only the localization of stereo sound images but also the psychoacoustic perception
of a spatial sound quality by a user.
For example, upon comparing two sound images which are generated using a time
panning method and an amplitude panning method, respectively, and reproduced at the
same location using a 2-channel stereo speaker, it is recognized that the amplitude
panning method can contribute to a precise localization of sound images, and that the
time panning method can provide natural sounds with a profound feeling of space.
Thus, if the renderer 133 only uses the amplitude panning method to arrange object
signals in a multi-channel space, the renderer 133 may be able to precisely localize
each sound image, but may not be able to provide as profound a feeling of sound as
when using the time panning method. Users may sometime prefer a precise lo-
calization of sound images to a profound feeling of sound or vice versa according to
the type of sound sources.
FIGS. 4(a) and 4(b) explains the influence of intensity (amplitude difference) and a
time difference on the localization of sound images as performed in the reproduction of
signals with a 2-channel stereo speaker. Referring to FIGS. 4(a) and 4(b), a sound
image may be localized at a predetermined angle according to an amplitude difference
and a time difference which are independent from each other. For example, an

amplitude difference of about 8 dB or a time difference of about 0.5 ms, which is
equivalent to the amplitude difference of 8 dB, may be used in order to localize a
sound image at an angle of 20 . Therefore, even if only an amplitude difference is
provided as mixing parameter information, it is possible to obtain various sounds with
different properties by converting the amplitude difference into a time difference
which is equivalent to the amplitude difference during the localization of sound
images.
FIG. 5 illustrates functions regarding the correspondence between amplitude
differences and time differences which are required to localize sound images at angles
of 10 , 20 , and 30 . The function illustrated in FIG. 5 may be obtained based on FIGS.
4(a) and 4(b). Referring to FIG. 5, various amplitude difference-time difference com-
binations may be provided for localizing a sound image at a predetermined position.
For example, assume that an amplitude difference of 8 dB is provided as mixing
parameter information in order to localize a sound image at an angle of 20 . According
to the function illustrated in FIG. 5, a sound image can also be localized at the angle of
20 using the combination of an amplitude difference of 3 dB and a time difference of
0.3 ms. In this case, not only amplitude difference information but also time difference
information may be provided as mixing parameter information, thereby enhancing the
feeling of space.
Therefore, in order to generate sounds with properties desired by a user during a
mixing/rendering operation, mixing parameter information may be appropriately
converted so that whichever of amplitude panning and time panning suits the user can
be performed. That is, if mixing parameter information only includes amplitude
difference information and the user wishes for sounds with a profound feeling of space,
the amplitude difference information may be converted into time difference in-
formation equivalent to the amplitude difference information with reference to psy-
choacoustic data. Alternatively, if the user wishes for both sounds with a profound
feeling of space and a precise localization of sound images, the amplitude difference
information may be converted into the combination of amplitude difference information
and time difference information equivalent to the original amplitude information.
Alternatively, if mixing parameter information only includes time
difference information and a user prefers a precise localization of sound images, the
time difference information may be converted into amplitude difference information
equivalent to the time difference information, or may be converted into the
combination of amplitude difference information and time difference information
which can satisfy the user's preference by enhancing both the precision of localization
of sound images and the feeling of space.
Still alternatively, if mixing parameter information includes both amplitude

difference information and time difference information and a user prefers a precise localization
of sound images, the combination of the amplitude difference information
and the time difference information may be converted into amplitude difference information
equivalent to the combination of the original amplitude difference information
and the time difference information. On the other hand, if mixing parameter
information includes both amplitude difference information and time difference information
and a user prefers the enhancement of the feeling of space, the combination
of the amplitude difference information and the time difference information may be
converted into time difference information equivalent the combination of the amplitude
difference information and the original time difference information. Referring to FIG.
6, control information may include mixing/rendering information and harmonic information
regarding one or more object signals. The harmonic information may
include at least one of pitch i nformation, fundamental frequency information, and
dominant frequency band information regarding one or more object signals, and descriptions
of the energy and spectrum of each sub-band of each of the object signals.
The harmonic information may be used to process an object signal during a rendering
operation because the resolution of a renderer which performs its operation in units of
sub-bands is insufficient.
If the harmonic information includes pitch information regarding one or more object
signals, the gain of each of the object signals may be adjusted by attenuating or
strengthening a predetermined frequency domain using a comb filter or an inverse
comb filter. For example, if one of a plurality of object signals is a vocal signal, the
object signals may be used as a karaoke by attenuating only the vocal signal. Alternatively,
if the harmonic information includes dominant frequency domain information
regarding one or more object signals, a process of attenuating or
strengthening a dominant frequency domain may be performed. Still alternatively, if
the harmonic information includes spectrum information regarding one or more object
signals, the gain of each of the object signals may be controlled by performing at-
tenuation or enforcement without being restricted by any sub-band boundaries.
FIG. 7 is a block diagram of an audio decoding apparatus 140 according to another
embodiment of the present invention. Referring to FIG. 7, the audio decoding
apparatus 140 uses a multi-channel decoder 141, instead of an object decoder and a
renderer, and decodes a number of object signals after the object signals are appropriately
arranged in a multi-channel space.
More specifically, the audio decoding apparatus 140 includes the multi-channel
decoder 141 and a parameter converter 145. The multi-channel decoder 141 generates
a multi-channel signal whose object signals have already been arranged in a multichannel
space based on a down-mix signal and spatial parameter information, which is

channel-based side information provided by the parameter converter 145. The
parameter converter 145 analyzes side information and control information transmitted
by an audio encoding apparatus (not shown), and generates the spatial parameter information
based on the result of the analysis. More specifically, the parameter
converter 145 generates the spatial parameter information by combining the side information
and the control information which includes playback setup information and
mixing information. That is, the parameter conversion 145 performs the conversion of
the combination of the side information and the control information to spatial data corresponding
to a One-To-Two (OTT) box or a Two-To-Three (TTT) box.
The audio decoding apparatus 140 may perform a multi-channel decoding operation
into which an object-based decoding operation and a mixing/rendering operation are
incorporated and may thus skip the decoding of each object signal. Therefore, it is
possible to reduce the complexity of decoding and/or mixing/rendering.
For example, when there are 10 object signals and a multi-channel signal obtained
based on the 10 object signals is to be reproduced by a 5.1 channel speaker reproduction
system, a typical object-based audio decoding apparatus generates decoded
signals respectively corresponding the 10 object signals based on a down-mix signal
and side information and then generates a 5.1 channel signal by appropriately
arranging the 10 object signals in a multi-channel space so that the object signals can
become suitable for a 5.1 channel speaker environment. However, it is inefficient to
generate 10 object signals during the generation of a 5.1 channel signal, and this
problem becomes more severe as the difference between the number of object signals
and the number of channels of a multi-channel signal to be generated increases.
On the other hand, according to the embodiment of FIG. 7, the audio decoding
apparatus 140 generates spatial parameter information suitable for a 5.1-channel signal
based on side information and control information, and provides the spatial parameter
information and a downmix signal to the multi-channel decoder 141. Then, the multichannel
decoder 141 generates a 5.1 channel signal based on the spatial parameter information
and the downmix signal. In other words, when the number of channels to be
output is 5.1 channels, the audio decoding apparatus 140 can readily generate a
5.1-channel signal based on a downmix signal without the need to generate 10 object
signals and is thus more efficient than a conventional audio decoding apparatus in
terms of complexity.
The audio decoding apparatus 140 is deemed efficient when the amount of
computation required to calculates spatial parameter information corresponding to each
of an OTT box and a TTT box through the analysis of side information and control information
transmitted by an audio encoding apparatus is less than the amount of
computation required to perform a mixing/rendering operation after the decoding of

each object signal.
The audio decoding apparatus 140 may be obtained simply by adding a module for
generating spatial parameter information through the analysis of side information and
control information to a typical multi-channel audio decoding apparatus, and may thus
maintain the compatibility with a typical multi-channel audio decoding apparatus.
Also, the audio decoding apparatus 140 can improve the quality of sound using
existing tools of a typical multi-channel audio decoding apparatus such as an envelope
shaper, a sub-band temporal processing (STP) tool, and a decorrelator. Given all this, it
is concluded that all the advantages of a typical multi-channel audio decoding method
can be readily applied to an object-audio decoding method.
Spatial parameter information transmitted to the multi-channel decoder 141 by the
parameter converter 145 may have been compressed so as to be suitable for being
transmitted. Alternatively, the spatial parameter information may have the same format
as that of data transmitted by a typical multi-channel encoding apparatus. That is, the
spatial parameter information may have been subjected to a Huffman decoding
operation or a pilot decoding operation and may thus be transmitted to each module as
uncompressed spatial cue data. The former is suitable for transmitting the spatial
parameter information to a multi-channel audio decoding apparatus in a remote place,
and the later is convenient because there is no need for a multi-channel audio decoding
apparatus to convert compressed spatial cue data into uncompressed spatial cue data
that can readily be used in a decoding operation.
The configuration of spatial parameter information based on the analysis of side information
and control information may cause a delay between a downmix signal and
the spatial parameter information. In order to address this, an additional buffer may be
provided either for a downmix signal or for spatial parameter information so that the
downmix signal and the spatial parameter information can be synchronized with each
other. These methods, however, are inconvenient because of the requirement to
provide an additional buffer. Alternatively, side information may be transmitted ahead
of a downmix signal in consideration of the possibility of occurrence of a delay
between a downmix signal and spatial parameter information. In this case, spatial
parameter information obtained by combining the side information and control information
does not need to be adjusted but can readily be used.
If a plurality of object signals of a downmix signal have different levels, an artistic
downmix gains (ADG) module which can directly compensate for the downmix signal
may determine the relative levels of the object signals, and each of the object signals
may be allocated to a predetermined position in a multi-channel space using spatial cue
data such as channel level difference information, inter-channel correlation (ICC) information,
and channel prediction coefficient (CPC) information.

For example, if control information indicates that a predetermined object signal is to
be allocated to a predetermined position in a multi-channel space and has a higher level
than other object signals, a typical multi-channel decoder may calculate the difference
between the energies of channels of a downmix signal, and divide the downmix signal
into a number of output channels based on the results of the calculation. However, a
typical multi-channel decoder cannot increase or reduce the volume of a certain sound
in a downmix signal. In other words, a typical multi-channel decoder simply distributes
a downmix signal to a number of output channels and thus cannot increase or reduce
the volume of a sound in the downmix signal.
It is relatively easy to allocate each of a number of object signals of a downmix
signal generated by an object encoder to a predetermined position in a multi-channel
space according to control information. However, special techniques are required to
increase or reduce the amplitude of a predetermined object signal. In other words, if a
downmix signal generated by an object encoder is used as it is, it is difficult to reduce
the amplitude of each object signal of the downmix signal.
Therefore, according to an embodiment of the present invention, the relative
amplitudes of object signals may be varied according to control information using an
ADG module 147 illustrated in FIG. 8. More specifically, the amplitude of any one of
a plurality of object signals of a downmix signal transmitted by an object encoder may
be increased or reduced using the ADG module 147. A downmix signal obtained by
compensation performed by the ADG module 147 may be subjected to multi-channel
decoding.
If the relative amplitudes of object signals of a downmix signal are appropriately
adjusted using the ADG module 147, it is possible to perform object decoding using a
typical multi-channel decoder. If a downmix signal generated by an object encoder is a
mono or stereo signal or a multi-channel signal with three or more channels, the
downmix signal may be processed by the ADG module 147. If a downmix signal
generated by an object encoder has two or more channels and a predetermined object
signal that needs to be adjusted by the ADG module 147 only exists in one of the
channels of the downmix signal, the ADG module 147 may be applied only to the
channel including the predetermined object signal, instead of being applied to all the
channels of the downmix signal. A downmix signal processed by the ADG module 147
in the above-described manner may be readily processed using a typical multi-channel
decoder without the need to modify the structure of the multi-channel decoder.
Even when a final output signal is not a multi-channel signal that can be reproduced
by a multi-channel speaker but is a binaural signal, the ADG module 147 may be used
to adjust the relative amplitudes of object signals of the final output signal.
Alternatively to the use of the ADG module 147, gain information specifying a gain

value to be applied to each object signal may be included in control information during
the generation of a number of object signals. For this, the structure of a typical multichannel
decoder may be modified. Even though requiring a modification to the
structure of an existing multi-channel decoder, this method is convenient in terms of
reducing the complexity of decoding by applying a gain value to each object signal
during a decoding operation without the need to calculate ADG and to compensate for
each object signal.
FIG. 9 is a block diagram of an audio decoding apparatus 150 according to a fourth
embodiment of the present invention. Referring to FIG. 9, the audio decoding
apparatus 150 is characterized by generating a binaural signal.
More specifically, the audio decoding apparatus 150 includes a multi-channel
binaural decoder 151, a first parameter converter 157, and a second parameter
converter 159.
The second parameter converter 159 analyzes side information and control information
which are provided by an audio encoding apparatus, and configures spatial
parameter information based on the result of the analysis. The first parameter converter
157 configures binaural parameter information, which can be used by the multichannel
binaural decoder 151, by adding three-dimensional (3D) information such as
head-related transfer function (HRTF) parameters to the spatial parameter information.
The multi-channel binaural decoder 151 generates a virtual three-dimensianl (3D)
signal by applying the virtual 3D parameter information to a downmix signal.
The first parameter converter 157 and the second parameter converter 159 may be
replaced by a single module, i.e., a parameter conversion module 155 which receives
the side information, the control information, and the HRTF parameters and configures
the binaural parameter information based on the side information, the control information,
and the HRTF parameters.
Conventionally, in order to generate a binaural signal for the reproduction of a
downmix signal including 10 object signals with a headphone, an object signal must
generate 10 decoded signals respectively corresponding to the 10 object signals based
on the downmix signal and side information. Thereafter, a Tenderer allocates each of
the 10 object signals to a predetermined position in a multi-channel space with
reference to control information so as to suit a 5-channel speaker environment.
Thereafter, the renderer generates a 5-channel signal that can be reproduced using a
5-channel speaker. Thereafter, the renderer applies HRTF parameters to the 5-channel
signal, thereby generating a 2-channel signal. In short, the above-mentioned conventional
audio decoding method includes reproducing 10 object signals, converting
the 10 object signals into a 5-channel signal, and generating a 2-channel signal based
on the 5-channel signal, and is thus inefficient.

On the other hand, the audio decoding apparatus 150 can readily generate a binaural
signal that can be reproduced using a headphone based on object audio signals. In
addition, the audio decoding apparatus 150 configures spatial parameter information
through the analysis of side information and control information, and can thus generate
a binaural signal using a typical multi-channel binaural decoder. Moreover, the audio
decoding apparatus 150 still can use a typical multi-channel binaural decoder even
when being equipped with an incorporated parameter converter which receives side information,
control information, and HRTF parameters and configures binaural
parameter information based on the side information, the control information, and the
HRTF parameters.
FIG. 10 is a block diagram of an audio decoding apparatus 160 according to a fifth
embodiment of the present invention. Referring to FIG. 10, the audio decoding
apparatus 160 includes a downmix processor 161, a multi-channel decoder 163, and a
parameter converter 165. The downmix processor 161 and the parameter converter 163
may be replaced by a single module 167.
The parameter converter 165 generates spatial parameter information, which can be
used by the multi-channel decoder 163, and parameter information, which can be used
by the downmix processor 161. The downmix processor 161 performs a pre-processing
operation on a downmix signal, and transmits a downmix signal resulting from the preprocessing
operation to the multi-channel decoder 163. The multi-channel decoder 163
performs a decoding operation on the downmix signal transmitted by the downmix
processor 161, thereby outputting a stereo signal, a binaural stereo signal or a multichannel
signal. Examples of the pre-processing operation performed by the downmix
processor 161 include the modification or conversion of a downmix signal in a time
domain or a frequency domain using filtering.
If a downmix signal input to the audio decoding apparatus 160 is a stereo signal, the
downmix signal may have be subjected to downmix preprocessing performed by the
downmix processor 161 before being input to the multi-channel decoder 163 because
the multi-channel decoder 163 cannot map a component of the downmix signal corresponding
to a left channel, which is one of multiple channels, to a right channel,
which is another of the multiple channels. Therefore, in order to shift the position of an
object signal classified into the left channel to the direction of the right channel, the
downmix signal input to the audio decoding apparatus 160 may be preprocessed by the
downmix processor 161, and the preprocessed downmix signal may be input to the
multi-channel decoder 163.
The preprocessing of a stereo downmix signal may be performed based on preprocessing
information obtained from side information and from control information.
FIG. 11 is a block diagram of an audio decoding apparatus 170 according to a sixth

embodiment of the present invention. Referring to FIG. 11, the audio decoding
apparatus 170 includes a multi-channel decoder 171, a channel processor 173, and a
parameter converter 175.
The parameter converter 175 generates spatial parameter information, which can be
used by the multi-channel decoder 173, and parameter information, which can be used
by the channel processor 173. The channel processor 173 performs a post-processing
operation on a signal output by the multi-channel decoder 173. Examples of the signal
output by the multi-channel decoder 173 include a stereo signal, a binaural stereo
signal and a multi-channel signal.
Examples of the post-processing operation performed by the post processor 173
include the modification and conversion of each channel or all channels of an output
signal. For example, if side information includes fundamental frequency information
regarding a predetermined object signal, the channel processor 173 may remove
harmonic components from the predetermined object signal with reference to the
fundamental frequency information. A multi-channel audio decoding method may not
be efficient enough to be used in a karaoke system. However, if fundamental frequency
information regarding vocal object signals is included in side information and
harmonic components of the vocal object signals are removed during a post-processing
operation, it is possible to realize a high-performance karaoke system using the
embodiment of FIG. 11. The embodiment of FIG. 11 may also be applied to object
signals, other than vocal object signals. For example, it is possible to remove the sound
of a predetermined musical instrument using the embodiment of FIG. 11. Also, it is
possible to amplify predetermined harmonic components using fundamental frequency
information regarding object signals using the embodiment of FIG. 11.
The channel processor 173 may perform additional effect processing on a downmix
signal. Alternatively, the channel processor 173 may add a signal obtained by the
additional effect processing to a signal output by the multi-channel decoder 171. The
channel processor 173 may change the spectrum of an object or modify a downmix
signal whenever necessary. If it is not appropriate to directly perform an effect
processing operation such as reverberation on a downmix signal and to transmit a
signal obtained by the effect processing operation to the multi-channel decoder 171,
the downmix processor 173 may add the signal obtained by the effect processing
operation to the output of the multi-channel decoder 171, instead of performing effect
processing on the downmix signal.
The audio decoding apparatus 170 may be designed to include not only the channel
processor 173 but also a downmix processor. In this case, the downmix processor may
be disposed in front of the multi-channel decoder 173, and the channel processor 173
may be disposed behind the multi-channel decoder 173.

FIG. 12 is a block diagram of an audio decoding apparatus 210 according to a
seventh embodiment of the present invention. Referring to FIG. 12, the audio decoding
apparatus 210 uses a multi-channel decoder 213, instead of an object decoder.
More specifically, the audio decoding apparatus 210 includes the multi-channel
decoder 213, a transcoder 215, a renderer 217, and a 3D information database 217.
The renderer 217 determines the 3D positions of a plurality of object signals based
on 3D information corresponding to index data included in control information. The
transcoder 215 generates channel-based side information by synthesizing position information
regarding a number of object audio signals to which 3D information is
applied by the renderer 217. The multi-channel decoder 213 outputs a 3D signal by
applying the channel-based side information to a down-mix signal
A head-related transfer function (HRTF) may be used as the 3D information. An
HRTF is a transfer function which describes the transmission of sound waves between
a sound source at an arbitrary position and the eardrum, and returns a value that varies
according to the direction and altitude of the sound source. If a signal with no directivity
is filtered using the HRTF, the signal may be heard as if it were reproduced
from a certain direction.
When an input bitstream is received, the audio decoding apparatus 210 extracts an
object-based downmix signal and object-based parameter information from the input
bitstream using a demultiplexer (not shown). Then, the renderer 217 extracts index
data from control information, which is used to determine the positions of a plurality of
object audio signals, and withdraws 3D information corresponding to the extracted
index data from the 3D information database 219.
More specifically, mixing parameter information, which is included in control information
that is used by the audio decoding apparatus 210, may include not only level
information but also index data necessary for searching for 3D information. The
mixing parameter information may also include time information regarding the time
difference between channels, position information and one or more parameters
obtained by appropriately combining the level information and the time information.
The position of an object audio signal may be determined initially according to
default mixing parameter information, and may be changed later by applying 3D information
corresponding to a position desired by a user to the object audio signal. Alternatively,
if the user wishes to apply a 3D effect only to several object audio signals,
level information and time information regarding other object audio signals to which
the user wishes not to apply a 3D effect may be used as mixing parameter information.
The transcoder 217 generates channel-based side information regarding M channels
by synthesizing object-based parameter information regarding N object signals
transmitted by an audio encoding apparatus and position information of a number of

object signals to which 3D information such as an HRTF is applied by the renderer
217.
The multi-channel decoder 213 generates an audio signal based on a downmix signal
and the channel-based side information provided by the transcoder 217, and generates
a 3D multi-channel signal by performing a 3D rendering operation using 3D information
included in the channel-based side information.
FIG. 13 is a block diagram of an audio decoding apparatus 220 according to a eighth
embodiment of the present invention. Referring to FIG. 13, the audio decoding
apparatus 220 is different from the audio decoding apparatus 210 illustrated in FIG. 12
in that a transcoder 225 transmits channel-based side information and 3D information
separately to a multi-channel decoder 223. In other words, the transcoder 225 of the
audio decoding apparatus 220 obtains channel-based side information regarding M
channels from object-based parameter information regarding N object signals and
transmits the channel-based side information and 3D information, which is applied to
each of the N object signals, to the multi-channel decoder 223, whereas the transcoder
217 of the audio decoding apparatus 210 transmits channel-based side information
including 3D information to the multi-channel decoder 213.
Referring to FIG. 14, channel-based side information and 3D information may
include a plurality of frame indexes. Thus, the multi-channel decoder 223 may
synchronize the channel-based side information and the 3D information with reference
to the frame indexes of each of the channel-based side information and the 3D information,
and may thus apply 3D information to a frame of a bitstream corresponding
to the 3D information. For example, 3D information having index 2 may be applied at
the beginning of frame 2 having index 2.
Since channel-based side information and 3D information both includes frame
indexes, it is possible to effectively determine a temporal position of the channel-based
side information to which the 3D information is to be applied, even if the 3D information
is updated over time. In other words, the transcoder 225 includes 3D information
and a number of frame indexes in channel-based side information, and thus,
the multi-channel decoder 223 can easily synchronize the channel-based side information
and the 3D information.
The downmix processor 231, transcoder 235, renderer 237 and the 3D information
database may be replaced by a single module 239.
FIG. 15 is a block diagram of an audio decoding apparatus 230 according to a ninth
embodiment of the present invention. Referring to FIG. 15, the audio decoding
apparatus 230 is differentiated from the audio decoding apparatus 220 illustrated in
FIG. 14 by further including a downmix processor 231.
More specifically, the audio decoding apparatus 230 includes a transcoder 235, a

Tenderer 237, a 3D information database 239, a multi-channel decoder 233, and the
downmix processor 231. The transcoder 235, the renderer 237, the 3D information
database 239, and the multi-channel decoder 233 are the same as their respective
counterparts illustrated in FIG. 14. The downmix processor 231 performs a preprocessing
operation on a stereo downmix signal for position adjustment. The 3D information
database 239 may be incorporated with the renderer 237. A module for
applying a predetermined effect to a downmix signal may also be provided in the audio
decoding apparatus 230.
FIG. 16 illustrates a block diagram of an audio decoding apparatus 240 according to
a tenth embodiment of the present invention. Referring to FIG. 16, the audio decoding
apparatus 240 is differentiated from the audio decoding apparatus 230 illustrated in
FIG. 15 by including a multi-point control unit combiner 241.
That is, the audio decoding apparatus 240, like the audio decoding apparatus 230,
includes a downmix processor 243, a multi-channel decoder 244, a transcoder 245, a
renderer 247, and a 3D information database 249. The multi-point control unit
combiner 241 combines a plurality of bitstreams obtained by object-based encoding,
thereby obtaining a single bitstream. For example, when a first bitstream for a first
audio signal and a second bitstream for a second audio signal are input, the multi-point
control unit combiner 241 extracts a first downmix signal from the first bitstream,
extracts a second downmix signal from the second bitstream and generates a third
downmix signal by combining the first and second downmix signals. In addition, the
multi-point control unit combiner 241 extracts first object-based side information from
the first bitstream, extract second object-based side information from the second
bitstream, and generates third object-based side information by combining the first
object-based side information and the second object-based side information.
Thereafter, the multi-point control unit combiner 241 generates a bitstream by
combining the third downmix signal and the third object-based side information and
outputs the generated bitstream.
Therefore, according to the tenth embodiment of the present invention, it is possible
to efficiently process even signals transmitted by two or more communication partners
compared to the case of encoding or decoding each object signal.
In order for the multi-point control unit combiner 241 to incorporate a plurality of
downmix signals, which are respectively extracted from a plurality of bitstreams and
are associated with different compression codecs, into a single downmix signal, the
downmix signals may need to be converted into pulse code modulation (PCM) signals
or signals in a predetermined frequency domain according to the types of the
compression codecs of the downmix signals, the PCM signals or the signals obtained
by the conversion may need to be combined together, and a signal obtained by the

combination may need to be converted using a predetermined compression codec. In
this case, a delay may occur according to whether the downmix signals are incorporated
into a PCM signal or into a signal in the predetermined frequency domain.
The delay, however, may not be able to be properly estimated by a decoder. Therefore,
the delay may need to be included in a bitstream and transmitted along with the
bitstream. The delay may indicate the number of delay samples in a PCM signal or the
number of delay samples in the predetermined frequency domain.
During an object-based audio coding operation, a considerable number of input
signals may sometimes need to be processed compared to the number of input signals
generally processed during a typical multi-channel coding operation (e.g., a
5.1-channel or 7.1-channel coding operation). Therefore, an object-based audio coding
method requires much higher bitrates than a typical channel-based multi-channel audio
coding method. However, since an object-based audio coding method involves the
processing of object signals which are smaller than channel signals, it is possible to
generate dynamic output signals using an object-based audio coding method.
An audio encoding method according to an embodiment of the present invention will
hereinafter be described in detail with reference to FIGS. 17 through 20.
In an object-based audio encoding method, object signals may be defined to
represent individual sounds such as the voice of a human or the sound of a musical
instrument. Alternatively, sounds having similar characteristics such as the sounds of
stringed musical instruments (e.g., a violin, a viola, and a cello), sounds belonging to
the same frequency band, or sounds classified into the same category according to the
directions and angles of their sound sources, may be grouped together, and defined by
the same object signals. Still alternatively, object signals may be defined using the
combination of the above-described methods.
A number of object signals may be transmitted as a downmix signal and side in-
formation. During the creation of information to be transmitted, the energy or power of
a downmix signal or each of a plurality of object signals of the downmix signal is
calculated originally for the purpose of detecting the envelope of the downmix signal.
The results of the calculation may be used to transmit the object signals or the
downmix signal or to calculate the ratio of the levels of the object signals.
A linear predictive coding (LPC) algorithm may be used to lower bitrates. More
specifically, a number of LPC coefficients which represent the envelope of a signal are
generated through the analysis of the signal, and the LPC coefficients are transmitted,
instead of transmitting envelop information regarding the signal. This method is
efficient in terms of bitrates. However, since the LPC coefficients are very likely to be
discrepant from the actual envelope of the signal, this method requires an addition
process such as error correction. In short, a method that involves transmitting envelop

information of a signal can guarantee a high quality of sound, but results in a considerable
increase in the amount of information that needs to be transmitted. On the
other hand, a method that involves the use of LPC coefficients can reduce the amount
of information that needs to be transmitted, but requires an additional process such as
error correction and results in a decrease in the quality of sound.
According to an embodiment of the present invention, a combination of these
methods may be used. In other words, the envelope of a signal may be represented by
the energy or power of the signal or an index value or another value such as an LPC
coefficient corresponding to the energy or power of the signal.
Envelope information regarding a signal may be obtained in units of temporal
sections or frequency sections. More specifically, referring to FIG. 17, envelope information
regarding a signal may be obtained in units of frames. Alternatively, if a
signal is represented by a frequency band structure using a filter bank such as a
quadrature mirror filter (QMF) bank, envelope information regarding a signal may be
obtained in units of frequency sub-bands, frequency sub-band partitions which are
smaller entities than frequency sub-bands, groups of frequency sub-bands or groups of
frequency sub-band partitions. Still alternatively, a combination of the frame-based
method, the frequency sub-band-based method, and the frequency sub-band partition-
based method may be used within the scope of the present invention.
Still alternatively, given that low-frequency components of a signal generally have
more information than high-frequency components of the signal, envelop information
regarding low-frequency components of a signal may be transmitted as it is, whereas
envelop information regarding high-frequency components of the signal may be
represented by LPC coefficients or other values and the LPC coefficients or the other
values may be transmitted instead of the envelop information regarding the high-
frequency components of the signal. However, low-frequency components of a signal
may not necessarily have more information than high-frequency components of the
signal. Therefore, the above-described method must be flexibly applied according to
the circumstances.
According to an embodiment of the present invention, envelope information or index
data corresponding to a portion (hereinafter referred to as the dominant portion) of a
signal that appears dominant on a time/frequency axis may be transmitted, and none of
envelope information and index data corresponding to a non-dominant portion of the
signal may be transmitted. Alternatively, values (e.g., LPC coefficients) that represent
the energy and power of the dominant portion of the signal may be transmitted, and no
such values corresponding to the non-dominant portion of the signal may be
transmitted. Still alternatively, envelope information or index data corresponding to the
dominant portion of the signal may be transmitted, and values that represent the energy

or power of the non-dominant portion of the signal may be transmitted. Still alternatively,
information only regarding the dominant portion of the signal may be
transmitted so that the non-dominant portion of the signal can be estimated based on
the information regarding the dominant portion of the signal. Still alternatively, a
combination of the above-described methods may be used.
For example, referring to FIG. 18, if a signal is divided into a dominant period and a
non-dominant period, information regarding the signal may be transmitted in four
different manners, as indicated by (a) through (d).
In order to transmit a number of object signals as the combination of a downmix
signal and side information, the downmix signal needs to be divided into a plurality of
elements as part of a decoding operation, for example, in consideration of the ratio of
the levels of the object signals. In order to guarantee independence between the
elements of the downmix signal, a decorrelation operation needs to be additionally
performed.
Object signals which are the units of coding in an object-based coding method have
more independence than channel signals which are the units of coding in a multichannel
coding method. In other words, a channel signal includes a number of object
signals, and thus needs to be decorrelated. On the other hand, object signals are independent
from one another, and thus, channel separation may be easily performed
simply using the characteristics of the object signals without a requirement of a
decorrelation operation.
More specifically, referring to FIG. 19, object signals A, B, and C take turns to
appear dominant on a frequency axis. In this case, there is no need to divide a
downmix signal into a number of signals according to the ratio of the levels of the
object signals A, B, and C and to perform decorrelation. Instead, information regarding
the dominant periods of the object signals A, B, and C may be transmitted, or a gain
value may be applied to each frequency component of each of the object signals A, B,
and C, thereby skipping decorrelation. Therefore, it is possible to reduce the amount of
computation and to reduce the bitrate by the amount that would have otherwise been
required by side information necessary for decorrelation.
In short, in order to skip decorrelation, which is performed so as to guarantee independence
among a number of signals obtained by dividing a downmix signal
according to the ratio of the ratios of object signals of the downmix signal, information
regarding a frequency domain including each object signal may be transmitted as side
information. Alternatively, different gain values may be applied to a dominant period
during which each object signal appears dominant and a non-dominant period during
which each object signal appears less dominant, and thus, information regarding the
dominant period may be mainly provided as side information. Still alternatively, the in-

formation regarding the dominant period may be transmitted as side information, and
no information regarding the non-dominant period may be transmitted. Still alternatively,
a combination of the above-described methods which are alternatives to a
decorrelation method may be used.
The above-described methods which are alternatives to a decorrelation method may
be applied to all object signals or only to some object signals with easily distinguishable
dominant periods. Also, the above-described methods which are alternatives
to a decorrelation method may be variably applied in units of frames.
The encoding of object audio signals using a residual signal will hereinafter be
described in detail.
In general, in an object-based audio coding method, a number of object signals are
encoded, and the results of the encoding are transmitted as the combination of a
downmix signal and side information. Then, a number of object signals are restored
from the downmix signal through decoding according to the side information, and the
restored object signals are appropriately mixed, for example, at the request of a user
according to control information, thereby generating a final channel signal. An object-
based audio coding method generally aims to freely vary an output channel signal
according to control information with the aid of a mixer. However, an object-based
audio coding method may also be used to generate a channel output in a predefined
manner regardless of control information.
For this, side information may include not only information necessary to obtain a
number of object signals from a downmix signal but also mixing parameter information
necessary to generate a channel signal. Thus, it is possible to generate a final
channel output signal without the aid of a mixer. In this case, such an algorithm as
residual coding may be used to improve the quality of sound.
A typical residual coding method includes coding a signal and coding the error
between the coded signal and the original signal, i.e., a residual signal. During a
decoding operation, the coded signal is decoded while compensating for the error
between the coded signal and the original signal, thereby restoring a signal that is as
similar to the original signal as possible. Since the error between the coded signal and
the original signal is generally inconsiderable, it is possible to reduce the amount of information
additionally necessary to perform residual coding.
If a final channel output of a decoder is fixed, not only mixing parameter information
necessary for generating a final channel signal but also residual coding information
may be provided as side information. In this case, it is possible to improve the quality
of sound.
FIG. 20 is a block diagram of an audio encoding apparatus 310 according to an
embodiment of the present invention. Referring to FIG. 20, the audio encoding

apparatus 310 is characterized by using a residual signal.
More specifically, the audio encoding apparatus 310 includes an encoder 311, a
decoder 313, a first mixer 315, a second mixer 319, an adder 317 and a bitstream
generator 321.
The first mixer 315 performs a mixing operation on an original signal, and the
second mixer 319 performs a mixing operation on a signal obtained by performing an
encoding operation and then a decoding operation on the original signal. The adder 317
calculates a residual signal between a signal output by the first mixer 315 and a signal
output by the second mixer 319. The bitstream generator 321 adds the residual signal
to side information and transmits the result of the addition. In this manner, it is
possible to enhance the quality of sound.
The calculation of a residual signal may be applied to all portions of a signal or only
for low-frequency portions of a signal. Alternatively, the calculation of a residual
signal may be variably applied only to frequency domains including dominant signals
on a frame-by-frame basis. Still alternatively, a combination of the above-described
methods may be used.
Since the amount of side information including residual signal information is much
greater than the amount of side information including no residual signal information,
the calculation of a residual signal may be applied only to some portions of a signal
that directly affect the quality of sound, thereby preventing an excessive increase in
bitrate. The present invention can be realized as computer-readable code written on a
computer-readable recording medium. The computer-readable recording medium may
be any type of recording device in which data is stored in a computer-readable manner.
Examples of the computer-readable recording medium include a ROM, a RAM, a CD-
ROM, a magnetic tape, a floppy disc, an optical data storage, and a carrier wave (e.g.,
data transmission through the Internet). The computer-readable recording medium can
be distributed over a plurality of computer systems connected to a network so that
computer-readable code is written thereto and executed therefrom in a decentralized
manner. Functional programs, code, and code segments needed for realizing the
present invention can be easily construed by one of ordinary skill in the art.
Industrial Applicability
As described above, according to the present invention, sound images are localized
for each object audio signal by benefiting from the advantages of object-based audio
encoding and decoding methods. Thus, it is possible to offer more realistic sounds
through the reproduction of object audio signals. In addition, the present invention may
be applied to interactive games, and may thus provide a user with a more realistic
virtual reality experience.

While the present invention has been particularly shown and described with reference
to exemplary embodiments thereof, it will be understood by those of ordinary skill in
the art that various changes in form and details may be made therein without departing
from the spirit and scope of the present invention as defined by the following claims.

Claims
[1] An audio decoding method comprising:
extracting a downmix signal and object-based side information from an input
audio signal;
generating rendering information based on input control information; and
generating channel-based side information based on the rendering information
and the object-based side information.
[2] The audio decoding method of claim 1, further comprising generating a multichannel
audio signal based on the downmix signal and the channel-based side information.
[3] The audio decoding method of claim 1, wherein the control information
comprises at least one of three-dimensional (3D) information, mixing information
and harmonic information for processing a predetermined object
signal.
[4] The audio decoding method of claim 1, wherein the generating the rendering information,
comprises converting time information included in the mixing information
into equivalent amplitude information in response to a user command.
[5] The audio decoding method of claim 3, wherein the generating the rendering information,
further comprises converting amplitude information included in the
mixing information into equivalent time information in response to a user
command.
[6] The audio decoding method of claim 3, wherein the harmonic information
comprises at least one of pitch information, fundamental frequency information
and dominant frequency information of the predetermined object signal.
[7] The audio decoding method of claim 6, further comprising adjusting gain of the predetermined object signal based on the harmonic information.
[8] The audio decoding method of claim 6, further comprising compensating for an object signal in a predetermined frequency band based on the harmonic information.
[9] The audio decoding method of claim 1, further comprising compensating for a delay between the spatial information and the downmix signal.
[10] An audio decoding apparatus comprising:
a demultiplexer which extracts a downmix signal and object-based side information
from an input audio signal;
a renderer which generates rendering information based on input control information;
and
a transcoder which generates channel-based side information based on the

rendering information and the object-based side information.
[11] The audio decoding apparatus of claim 10, further comprising a multi-channel decoder which generates a multi-channel audio signal based on the downmix signal and the channel-based side information.
[12] The audio decoding apparatus of claim 10, wherein the control data comprises at least one of 3D information, mixing information and harmonic information for processing a predetermined object signal.
[13] The audio decoding apparatus of claim 12, wherein the renderer converts time information included in the mixing information into equivalent amplitude information in response to a user command during the generation of the rendering information.
[14] The audio decoding apparatus of claim 12, wherein the renderer converts
amplitude information included in the mixing information into equivalent time
information in response to a user command during the generation of the
rendering information.
[15] The audio decoding apparatus of claim 12, wherein the harmonic information comprises at least one of pitch information, fundamental frequency information and dominant frequency information of the predetermined object signal.
[16] The audio decoding apparatus of claim 15, wherein the renderer adjusts gain of the predetermined object signal based on the harmonic information.
[17] The audio decoding apparatus of claim 15, wherein the renderer compensates for an object signal in a predetermined frequency band based on the harmonic information.
[18] The audio decoding apparatus of claim 10, further comprising a buffer which
compensates for a delay between the spatial information and the downmix signal.
[19] A computer-readable recording medium having recorded thereon a computer
program for executing an audio decoding method, the audio decoding method
comprising:
extracting a downmix signal and object-based side information from an input
audio signal;
generating rendering information based on input control data; and
generating channel-based side information based on the rendering information
and the object-based side information.
[20] The computer-readable recording medium of claim 19, wherein the audio
decoding method further comprises generating a multi-channel audio signal
using the downmix signal and the channel-based side information.

Provided are an audio encoding method and apparatus and an audio decoding method and apparatus in which audio signals can be encoded or decoded so that sound images can be localized at any desired position for each object audio signal. The audio decoding method includes extracting a downmix signal and object-based side information from an input audio signal; generating rendering information based on input control data; and generating spatial information based on the rendering information and the object-based side information.

Documents:

http://ipindiaonline.gov.in/patentsearch/GrantedSearch/viewdoc.aspx?id=1wKH1QVZaorl1I8osvgb/g==&loc=wDBSZCsAt7zoiVrqcFJsRw==


Patent Number 272229
Indian Patent Application Number 3790/KOLNP/2008
PG Journal Number 14/2016
Publication Date 01-Apr-2016
Grant Date 22-Mar-2016
Date of Filing 17-Sep-2008
Name of Patentee LG ELECTRONICS INC.
Applicant Address 20, YEOEUIDO-DONG, YONGDEUNGPO-KU, SEOUL
Inventors:
# Inventor's Name Inventor's Address
1 PANG, HEE SUK #16 WOOMYUN-DONG, SEOCHO-KU, SEOUL 137-724
2 YOON, SUNG YONG #16 WOOMYUN-DONG, SEOCHO-KU, SEOUL 137-724
3 LEE, HYUN KOOK #16 WOOMYUN-DONG, SEOCHO-KU, GOYANG-SI, KYUNGGI-DO 137-724
4 KIM, DONG SOO #16 WOOMYUN-DONG, SEOCHO-KU, SEOUL 137-724
5 LIM, JAE HYUN #16 WOOMYUN-DONG, SEOCHO-KU, SEOUL 137-724
PCT International Classification Number G10L 19/00
PCT International Application Number PCT/KR2007/004801
PCT International Filing date 2007-10-01
PCT Conventions:
# PCT Application Number Date of Convention Priority Country
1 60/829,800 2006-10-17 U.S.A.
2 60/848,293 2006-09-29 U.S.A.
3 60/860,823 2006-11-24 U.S.A.
4 60/863,303 2006-10-27 U.S.A.
5 60/880,942 2007-01-18 U.S.A.