Title of Invention

APPARATUS AND METHOD FOR DECODING A SIGNAL

Abstract An encoding method and apparatus and a decoding method and apparatus are provided. The decoding method includes extracting a three-dimensional (3D) down-mix signal and spatial information from an input bitstream, removing 3D effects from the 3D down-mix signal by performing a 3D rendering operation on the 3D down-mix signal, and generating a multi-channel signal using the spatial information and a down-mix signal obtained by the removal. Accordingly, it is possible to efficiently encode multi-channel signals with 3D effects and to adaptively restore and reproduce audio signals with optimum sound quality according to the characteristics of a reproduction environment.
Full Text WO 2007/091842 PCT/KR2007/000668
Description
APPARATUS AND METHOD FOR ENCODING/DECODING
SIGNAL
Technical Field
The present invention relates to an encoding/decoding method and an encoding/
decoding apparatus, and more particularly, to an encoding/decoding apparatus which
can process an audio signal so that three dimensional (3D) sound effects can be
created, and an encoding/decoding method using the encoding/decoding apparatus.
Background Art
An encoding apparatus down-mixes a multi-channel signal into a signal with fewer
channels, and transmits the down-mixed signal to a decoding apparatus. Then, the
decoding apparatus restores a multi-channel signal from the down-mixed signal and
reproduces the restored multi-channel signal using three or more speakers, for
example, 5.1-channel speakers.
Multi-channel signals may be reproduced by 2-channel speakers such as
headphones. In this case, in order to make a user feel as if sounds output by 2-channel
speakers, were reproduced from three or more sound sources, it is necessary to develop
three-dimensional (3D) processing techniques capable of encoding or decoding multi-
channel signals so that 3D effects can be created.
Disclosure of Invention
Technical Problem
The present invention provides an encoding/decoding apparatus and an encoding/
decoding method which can reproduce multi-channel signals in various reproduction
environments by efficiently processing signals with 3D effects.
Technical Solution
According to an aspect of the present invention, there is provided a decoding
method of restoring a multi-channel signal, the decoding method including extracting a
three-dimensional (3D) down-mix signal and spatial information from an input
bitstream, removing 3D effects from the 3D down-mix signal by performing a 3D
rendering operation on the 3D down-mix signal, and generating a multi-channel signal
using the spatial information and a down-mix signal obtained by the removal.
According to another aspect of the present invention, there is provided a decoding
method of restoring a multi-channel signal, the decoding method including extracting a
3D down-mix signal and spatial information from an input bitstream, generating a
multi-channel signal using the 3D down-mix signal and the spatial information, and
removing 3D effects from the multi-channel signal by performing a 3D rendering

WO 2007/091842

PCT/KR2007/000668

operation on the multi-channel signal.
According to another aspect of the present invention, there is provided an encoding
method of encoding a multi-channel signal with a plurality of channels, the encoding
method including encoding the multi-channel signal into a down-mix signal with fewer
channels, generating spatial information regarding the plurality of channels, generating
a 3D down-mix signal by performing a 3D rendering operation on the down-mix
signal, and generating a bitstream including the 3D down-mix signal and the spatial in-
formation.
According to another aspect of the present invention, there is provided an encoding
method of encoding a multi-channel signal with a plurality of channels, the encoding
method including performing a 3D rendering operation on the multi-channel signal,
encoding a multi-channel signal obtained by the 3D rendering operation into a 3D
down-mix signal with fewer channels, generating spatial information regarding the
plurality of channels, and generating a bitstream including the 3D down-mix signal and
the spatial information.
According to another aspect of the present invention, there is provided a decoding
apparatus for restoring a multi-channel signal, the decoding apparatus including a bit
unpacking unit which extracts an encoded 3D down-mix signal and spatial information
from an input bitstream, a down-mix decoder which decodes the encoded 3D down-
mix signal, a 3D rendering unit which removes 3D effects from the decoded 3D down-
mix signal obtained by the decoding performed by the down-mix decoder by
performing a 3D rendering operation on the decoded 3D down-mix signal, and a multi-
channel decoder which generates a multi-channel signal using the spatial information
and a down-mix signal obtained by the removal performed by the 3D rendering unit.
According to another aspect of the present invention, there is provided a decoding
apparatus for restoring a multi-channel signal, the decoding apparatus including a bit
unpacking unit which extracts an encoded 3D down-mix signal and spatial information
from an input bitstream, a down-mix decoder which decodes the encoded 3D down-
mix signal, a multi-channel decoder which generates a multi-channel signal using the
spatial information and a 3D down-mix signal obtained by the decoding performed by
the down-mix decoder, and a 3D rendering unit which removes 3D effects from the
multi-channel signal by performing a 3D rendering operation on the multi-channel
signal.
According to another aspect of the present invention, there is provided an encoding
apparatus for encoding a multi-channel signal with a plurality of channels, the
encoding apparatus including a multi-channel encoder which encodes the multi-
channel signal into a down-mix signal with fewer channels and generates spatial in-
formation regarding the plurality of channels, a 3D rendering unit which generates a

WO 2007/091842 PCT7KR2007/000668
3D down-mix signal by performing a 3D rendering operation on the down-mix signal,
a down-mix encoder which encodes the 3D down-mix signal; and a bit packing unit
which generates a bitstream including the encoded 3D down-mix signal and the spatial
information.
According to another aspect of the present invention, there is provided an encoding
apparatus for encoding a multi-channel signal with a plurality of channels, the
encoding apparatus including a 3D rendering unit which performs a 3D rendering
operation on the multi-channel signal, a multi-channel encoder which encodes a multi-
channel signal obtained by the 3D rendering operation into a 3D down-mix signal with
fewer channels and generates spatial information regarding the plurality of channels, a
down-mix encoder which encodes the 3D down-mix signal, and a bit packing unit
which generates a bitstream including the encoded 3D down-mix signal and the spatial
information.
According to another aspect of the present invention, there is provided a bitstream
including a data field which includes information regarding a 3D down-mix signal, a
filter information field which includes filter information identifying a filter used for
generating the 3D down-mix signal, a first header field which includes information
indicating whether the filter information field includes the filter information, a second
header field which includes information indicating whether the filter information field
includes coefficients of the filter or coefficients of an inverse filter of the filter, and a
spatial information field which includes spatial information regarding a plurality of
channels.
According to another aspect of the present invention, there is provided a computer-
readable recording medium having a computer program for executing any one of the
above-described decoding methods and the above-described encoding methods.
Advantageous Effects
According to the present invention, it is possible to efficiently encode multi-channel
signals with 3D effects and to adaptively restore and reproduce audio signals with
optimum sound quality according to the characteristics of a reproduction environment.
Brief Description of the Drawings
FIG. 1 is a block diagram of an encoding/decoding apparatus according to an
embodiment of the present invention;
FIG. 2 is a block diagram of an encoding apparatus according to an embodiment of
the present invention;
FIG. 3 is a block diagram of a decoding apparatus according to an embodiment of
the present invention;
FIG. 4 is a block diagram of an encoding apparatus according to another

WO 2007/091842 PCT/KR2007/000668
embodiment of the present invention;
[20] FIG. 5 is a block diagram of a decoding apparatus according to another
embodiment of the present invention;
[21] FIG. 6 is a block diagram of a decoding apparatus according to another
embodiment of the present invention;
[22] FIG. 7 is a block diagram of a three-dimensional (3D) rendering apparatus
according to an embodiment of the present invention;
[23] FIGS. 8 through 11 illustrate bitstreams according to embodiments of the present
invention;
[24] FIG. 12 is a block diagram of an encoding/decoding apparatus for processing an
arbitrary down-mix signal according to an embodiment of the present invention;
[25] FIG. 13 is a block diagram of an arbitrary down-mix signal compensation/3D
rendering unit according to an embodiment of the present invention;
[26] FIG. 14 is a block diagram of a decoding apparatus for processing a compatible
down-mix signal according to an embodiment of the present invention;
[27] FIG. 15 is a block diagram of a down-mix compatibility processing/3D rendering
unit according to an embodiment of the present invention; and
[28] FIG. 16 is a block diagram of a decoding apparatus for canceling crosstalk
according to an embodiment of the present invention.
Best Mode for Carrying Out the Invention
[29] The present invention will hereinafter be described more fully with reference to the
accompanying drawings, in which exemplary embodiments of the invention are shown.
[30] FIG. 1 is a block diagram of an encoding/decoding apparatus according to an
embodiment of the present invention. Referring to FIG. 1, an encoding unit 100
includes a multi-channel encoder 110, a three-dimensional (3D) rendering unit 120, a
down-mix encoder 130, and a bit packing unit 140.
[31] The multi-channel encoder 110 down-mixes a multi-channel signal with a plurality
of channels into a down-mix signal such as a stereo signal or a mono signal and
generates spatial information regarding the channels of the multi-channel signal. The
spatial information is needed to restore a multi-channel signal from the down-mix
signal.
[32] Examples of the spatial information include a channel level difference (CLD),
which indicates the difference between the energy levels of a pair of channels, a
channel prediction coefficient (CPC), which is a prediction coefficient used to generate
a 3-channel signal based on a 2-channel signal, inter-channel correlation (ICC), which
indicates the correlation between a pair of channels, and a channel time difference
(CTD), which is the time interval between a pair of channels.

WO 2007/091842 PCT/KR2007/000668
The 3D rendering unit 120 generates a 3D down-mix signal based on the down-mix
signal. The 3D down-mix signal may be a 2-channel signal with three or more di-
rectivities and can thus be reproduced by 2-channel speakers such as headphones with
3D effects. In other words, the 3D down-mix signal may be reproduced by 2-channel
speakers so that a user can feel as if the 3D down-mix signal were reproduced from a
sound source with three or more channels. The direction of a sound source may be
determined based on at least one of the difference between the intensities of two
sounds respectively input to both ears, the time interval between the two sounds, and
the difference between the phases of the two sounds. Therefore, the 3D rendering unit
120 can convert the down-mix signal into the 3D down-mix signal based on how the
humans can determine the 3D location of a sound source with their sense of hearing.
The 3D rendering unit 120 may generate the 3D down-mix signal by filtering the
down-mix signal using a filter. In this case, filter-related information, for example, a
coefficient of the filter, may be input to the 3D rendering unit 120 by an external
source. The 3D rendering unit 120 may use the spatial information provided by the
multi-channel encoder 110 to generate the 3D down-mix signal based on the down-mix
signal. More specifically, the 3D rendering unit 120 may convert the down-mix signal
into the 3D down-mix signal by converting the down-mix signal into an imaginary
multi-channel signal using the spatial information and filtering the imaginary multi-
channel signal.
The 3D rendering unit 120 may generate the 3D down-mix signal by filtering the
down-mix signal using a head-related transfer function (HRTF) filter.
A HRTF is a transfer function which describes the transmission of sound waves
between a sound source at an arbitrary location and the eardrum, and returns a value
that varies according to the direction and altitude of a sound source. If a signal with no
directivity is filtered using the HRTF, the signal may be heard as if it were reproduced
from a certain direction.
The 3D rendering unit 120 may perform a 3D rendering operation in a frequency
domain, for example, a discrete Fourier transform (DFT) domain or a fast Fourier
transform (FFT) domain. In this case, the 3D rendering unit 120 may perform DFT or
FFT before the 3D rendering operation or may perform inverse DFT (IDFT) or inverse
FFT (IFFT) after the 3D rendering operation.
The 3D rendering unit 120 may perform the 3D rendering operation in a quadrature
mirror filter (QMF)/hybrid domain. In this case, the 3D rendering unit 120 may
perform QMF/hybrid analysis and synthesis operations before or after the 3D
rendering operation.
The 3D rendering unit 120 may perform the 3D rendering operation in a time
domain. The 3D rendering unit 120 may determine in which domain the 3D rendering

WO 2007/091842 PCT/KR2007/000668
operation is to be performed according to required sound quality and the operational
capacity of the encoding/decoding apparatus.
The down-mix encoder 130 encodes the down-mix signal output by the multi-
channel encoder 110 or the 3D down-mix signal output by the 3D rendering unit 120.
The down-mix encoder 130 may encode the down-mix signal output by the multi-
channel encoder 110 or the 3D down-mix signal output by the 3D rendering unit 120
using an audio encoding method such as an advanced audio coding (AAC) method, an
MPEG layer 3 (MP3) method, or a bit sliced arithmetic coding (BSAC) method.
The down-mix encoder 130 may encode a non-3D down-mix signal or a 3D down-
mix signal. In this case, the encoded non-3D down-mix signal and the encoded 3D
down-mix signal may both be included in a bitstream to be transmitted.
The bit packing unit 140 generates a bitstream based on the spatial information and
either the encoded non-3D down-mix signal or the encoded 3D down-mix signal.
The bitstream generated by the bit packing unit 140 may include spatial in-
formation, down-mix identification information indicating whether a down-mix signal
included in the bitstream is a non-3D down-mix signal or a 3D down-mix signal, and
information identifying a filter used by the 3D rendering unit 120 (e.g., HRTF co-
efficient information).
In other words, the bitstream generated by the bit packing unit 140 may include at
least one of a non-3D down-mix signal which has not yet been 3D-processed and an
encoder 3D down-mix signal which is obtained by a 3D processing operation
performed by an encoding apparatus, and down-mix identification information
identifying the type of down-mix signal included in the bitstream.
It may be determined which of the non-3D down-mix signal and the encoder 3D
down-mix signal is to be included in the bitstream generated by the bit packing unit
140 at the user's choice or according to the capabilities of the encoding/decoding
apparatus illustrated in FIG. 1 and the characteristics of a reproduction environment.
The HRTF coefficient information may include coefficients of an inverse function
of a HRTF used by the 3D rendering unit 120. The HRTF coefficient information may
only include brief information of coefficients of the HRTF used by the 3D rendering
unit 120, for example, envelope information of the HRTF coefficients. If a bitstream
including the coefficients of the inverse function of the HRTF is transmitted to a
decoding apparatus, the decoding apparatus does not need to perform an HRTF co-
efficient conversion operation, and thus, the amount of computation of the decoding
apparatus may be reduced.
The bitstream generated by the bit packing unit 140 may also include information
regarding an energy variation in a signal caused by HRTF-based filtering, i.e., in-
formation regarding the difference between the energy of a signal to be filtered and the

WO 2007/091842 PCT7KR2007/000668
energy of a signal that has been filtered or the ratio of the energy of the signal to be
filtered and the energy of the signal that has been filtered.
The bitstream generated by the bit packing unit 140 may also include information
indicating whether it includes HRTF coefficients. If HRTF coefficients are included in
the bitstream generated by the bit packing unit 140, the bitstream may also include in-
formation indicating whether it includes either the coefficients of the HRTF used by
the 3D rendering unit 120 or the coefficients of the inverse function of the HRTF.
Referring to FIG. 1, a first decoding unit 200 includes a bit unpacking unit 210, a
down-mix decoder 220, a 3D rendering unit 230, and a multi-channel decoder 240.
The bit unpacking unit 210 receives an input bitstream from the encoding unit 100
and extracts an encoded down-mix signal and spatial information from the input
bitstream. The down-mix decoder 220 decodes the encoded down-mix signal. The
down-mix decoder 220 may decode the encoded down-mix signal using an audio
signal decoding method such as an AAC method, an MP3 method, or a BSAC method.
As described above, the encoded down-mix signal extracted from the input
bitstream may be an encoded non-3D down-mix signal or an encoded, encoder 3D
down-mix signal. Information indicating whether the encoded down-mix signal
extracted from the input bitstream is an encoded non-3D down-mix signal or an
encoded, encoder 3D down-mix signal may be included in the input bitstream.
If the encoded down-mix signal extracted from the input bitstream is an encoder
3D down-mix signal, the encoded down-mix signal may be readily reproduced after
being decoded by the down-mix decoder 220.
On the other hand, if the encoded down-mix signal extracted from the input
bitstream is a non-3D down-mix signal, the encoded down-mix signal may be decoded
by the down-mix decoder 220, and a down-mix signal obtained by the decoding may
be converted into a decoder 3D down-mix signal by a 3D rendering operation
performed by the third rendering unit 233. The decoder 3D down-mix signal can be
readily reproduced.
The 3D rendering unit 230 includes a first renderer 231, a second Tenderer 232, and
a third renderer 233. The first renderer 231 generates a down-mix signal by performing
a 3D rendering operation on an encoder 3D down-mix signal provided by the down-
mix decoder 220. For example, the first renderer 231 may generate a non-3D down-
mix signal by removing 3D effects from the encoder 3D down-mix signal. The 3D
effects of the encoder 3D down-mix signal may not be completely removed by the first
renderer 231. In this case, a down-mix signal output by the first renderer 231 may have
some 3D effects.
The first renderer 231 may convert the 3D down-mix signal provided by the down-
mix decoder 220 into a down-mix signal with 3D effects removed therefrom using an

WO 2007/091842 PCT7KR2007/000668
inverse filter of the filter used by the 3D rendering unit 120 of the encoding unit 100.
Information regarding the filter used by the 3D rendering unit 120 or the inverse filter
of the filter used by the 3D rendering unit 120 may be included in the input bitstream.
The filter used by the 3D rendering unit 120 may be an HRTF filter. In this case,
the coefficients of the HRTF used by the encoding unit 100 or the coefficients of the
inverse function of the HRTF may also be included in the input bitstream. If the co-
efficients of the HRTF used by the encoding unit 100 are included in the input
bitstream, the HRTF coefficients may be inversely converted, and the results of the
inverse conversion may be used during the 3D rendering operation performed by the
first Tenderer 231. If the coefficients of the inverse function of the HRTF used by the
encoding unit 100 are included in the input bitstream, they may be readily used during
the 3D rendering operation performed by the first Tenderer 231 without being subjected
to any inverse conversion operation. In this case, the amount of computation of the first
decoding apparatus 100 may be reduced.
The input bitstream may also include filter information (e.g., information
indicating whether the coefficients of the HRTF used by the encoding unit 100 are
included in the input bitstream) and information indicating whether the filter in-
formation has been inversely converted.
The multi-channel decoder 240 generates a 3D multi-channel signal with three or
more channels based on the down-mix signal with 3D effects removed therefrom and
the spatial information extracted from the input bitstream.
The second Tenderer 232 may generate a 3D down-mix signal with 3D effects by
performing a 3D rendering operation on the down-mix signal with 3D effects removed
therefrom. In other words, the first Tenderer 231 removes 3D effects from the encoder
3D down-mix signal provided by the down-mix decoder 220. Thereafter, the second
renderer 232 may generate a combined 3D down-mix signal with 3D effects desired by
the first decoding apparatus 200 by performing a 3D rendering operation on a down-
mix signal obtained by the removal performed by the first Tenderer 231, using a filter
of the first decoding apparatus 200.
The first decoding apparatus 200 may include a renderer in which two or more of
the first, second, and third Tenderers 231, 232, and 233 that perform the same
operations are integrated.
A bitstream generated by the encoding unit 100 may be input to a second decoding
apparatus 300 which has a different structure from the first decoding apparatus 200.
The second decoding apparatus 300 may generate a 3D down-mix signal based on a
down-mix signal included in the bitstream input thereto.
More specifically, the second decoding apparatus 300 includes a bit unpacking unit
310, a down-mix decoder 320, and a 3D rendering unit 330. The bit unpacking unit

WO 2007/091842 PCT7KR2007/000668
310 receives an input bitstream from the encoding unit 100 and extracts an encoded
down-mix signal and spatial information from the input bitstream. The down-mix
decoder 320 decodes the encoded down-mix signal. The 3D rendering unit 330
performs a 3D rendering operation on the decoded down-mix signal so that the
decoded down-mix signal can be converted into a 3D down-mix signal.
FIG. 2 is a block diagram of an encoding apparatus according to an embodiment of
the present invention. Referring to FIG. 2, the encoding apparatus includes rendering
units 400 and 420 and a multi-channel encoder 410. Detailed descriptions of the same
encoding processes as those of the embodiment of FIG. 1 will be omitted.
Referring to FIG. 2, the 3D rendering units 400 and 420 may be respectively
disposed in front of and behind the multi-channel encoder 410. Thus, a multi-channel
signal may be 3D-rendered by the 3D rendering unit 400, and then, the 3D-rendered
multi-channel signal may be encoded by the multi-channel encoder 410, thereby
generating a pre-processed, encoder 3D down-mix signal. Alternatively, the multi-
channel signal may be down-mixed by the multi-channel encoder 410, and then, the
down-mixed signal may be 3D-rendered by the 3D rendering unit 420, thereby
generating a post-processed, encoder down-mix signal.
Information indicating whether the multi-channel signal has been 3D-rendered
before or after being down-mixed may be included in a bitstream to be transmitted.
The 3D rendering units 400 and 420 may both be disposed in front of or behind the
multi-channel encoder 410.
FIG. 3 is a block diagram of a decoding apparatus according to an embodiment of
the present invention. Referring to FIG. 3, the decoding apparatus includes 3D
rendering units 430 and 450 and a multi-channel decoder 440. Detailed descriptions of
the same decoding processes as those of the embodiment of FIG. 1 will be omitted.
Referring to FIG. 3, the 3D rendering units 430 and 450 may be respectively
disposed in front of and behind the multi-channel decoder 440. The 3D rendering unit
430 may remove 3D effects from an encoder 3D down-mix signal and input a down-
mix signal obtained by the removal to the multi-channel decoder 430. Then, the multi-
channel decoder 430 may decode the down-mix signal input thereto, thereby
generating a pre-processed 3D multi-channel signal. Alternatively, the multi-channel
decoder 430 may restore a multi-channel signal from an encoded 3D down-mix signal,
and the 3D rendering unit 450 may remove 3D effects from the restored multi-channel
signal, thereby generating a post-processed 3D multi-channel signal.
If an encoder 3D down-mix signal provided by an encoding apparatus has been
generated by performing a 3D rendering operation and then a down-mixing operation,
the encoder 3D down-mix signal may be decoded by performing a multi-channel
decoding operation and then a 3D rendering operation. On the other hand, if the

WO 2007/091842 PCT7KR2007/000668
encoder 3D down-mix signal has been generated by performing a down-mixing
operation and then a 3D rendering operation, the encoder 3D down-mix signal may be
decoded by performing a 3D rendering operation and then a multi-channel decoding
operation.
[70] Information indicating whether an encoded 3D down-mix signal has been obtained
by performing a 3D rendering operation before or after a down-mixing operation may
be extracted from a bitstream transmitted by an encoding apparatus.
[71] The 3D rendering units 430 and 450 may both be disposed in front of or behind the
multi-channel decoder 440.
[72] FIG,, 4 is a block diagram of an encoding apparatus according to another
embodiment of the present invention. Referring to FIG. 4, the encoding apparatus
includes a multi-channel encoder 500, a 3D rendering unit 510, a down-mix encoder
520, and a bit packing unit 530. Detailed descriptions of the same encoding processes
as those of the embodiment of FIG. 1 will be omitted.
[73] Referring to FIG. 4, the multi-channel encoder 500 generates a down-mix signal
and spatial information based on an input multi-channel signal. The 3D rendering unit
510 generates a 3D down-mix signal by performing a 3D rendering operation on the
down-mix signal.
[74] It may be determined whether to perform a 3D rendering operation on the down-
mix signal at a user's choice or according to the capabilities of the encoding apparatus,
the characteristics of a reproduction environment, or required sound quality.
[75] The down-mix encoder 520 encodes the down-mix signal generated by the multi-
channel encoder 500 or the 3D down-mix signal generated by the 3D rendering unit
510.
[76] The bit packing unit 530 generates a bitstream based on the spatial information and
either the encoded down-mix signal or an encoded, encoder 3D down-mix signal. The
bitstream generated by the bit packing unit 530 may include down-mix identification
information indicating whether an encoded down-mix signal included in the bitstream
is a non-3D down-mix signal with no 3D effects or an encoder 3D down-mix signal
with 3D effects. More specifically, the down-mix identification information may
indicate whether the bitstream generated by the bit packing unit 530 includes a non-3D
down-mix signal, an encoder 3D down-mix signal or both.
[77] FIG. 5 is a block diagram of a decoding apparatus according to another
embodiment of the present invention. Referring to FIG. 5, the decoding apparatus
includes a bit unpacking unit 540, a down-mix decoder 550, and a 3D rendering unit
560. Detailed descriptions of the same decoding processes as those of the embodiment
of FIG. 1 will be omitted.
[78] Referring to FIG. 5, the bit unpacking unit 540 extracts an encoded down-mix

WO 2007/091842 PCT/KR2007/000668
signal, spatial information, and down-mix identification information from an input
bitstream. The down-mix identification information indicates whether the encoded
down-mix signal is an encoded non-3D down-mix signal with no 3D effects or an
encoded 3D down-mix signal with 3D effects.
[79] If the input bitstream includes both a non-3D down-mix signal and a 3D down-mix
signal, only one of the non-3D down-mix signal and the 3D down-mix signal may be
extracted from the input bitstream at a user's choice or according to the capabilities of
the decoding apparatus, the characteristics of a reproduction environment or required
sound quality.
[80] The down-mix decoder 550 decodes the encoded down-mix signal. If a down-mix
signal obtained by the decoding performed by the down-mix decoder 550 is an encoder
3D down-mix signal obtained by performing a 3D rendering operation, the down-mix
signal may be readily reproduced.
[81] On the other hand, if the down-mix signal obtained by the decoding performed by
the down-mix decoder 550 is a down-mix signal with no 3D effects, the 3D rendering
unit 560 may generate a decoder 3D down-mix signal by performing a 3D rendering
operation on the down-mix signal obtained by the decoding performed by the down-
mix decoder 550.
[82] FIG. 6 is a block diagram of a decoding apparatus according to another
embodiment of the present invention. Referring to FIG. 6, the decoding apparatus
includes a bit unpacking unit 600, a down-mix decoder 610, a first 3D rendering unit
620, a second 3D rendering unit 630, and a filter information storage unit 640. Detailed
descriptions of the same decoding processes as those of the embodiment of FIG. 1 will
be omitted.
[83] The bit unpacking unit 600 extracts an encoded, encoder 3D down-mix signal and
spatial information from an input bitstream. The down-mix decoder 610 decodes the
encoded, encoder 3D down-mix signal.
[84] The first 3D rendering unit 620 removes 3D effects from an encoder 3D down-mix
signal obtained by the decoding performed by the down-mix decoder 610, using an
inverse filter of a filter of an encoding apparatus used for performing a 3D rendering
operation. The second rendering unit 630 generates a combined 3D down-mix signal
with 3D effects by performing a 3D rendering operation on a down-mix signal
obtained by the removal performed by the first 3D rendering unit 620, using a filter
stored in the decoding apparatus.
[85] The second 3D rendering unit 630 may perform a 3D rendering operation using a
filter having different characteristics from the filter of the encoding unit used to
perform a 3D rendering operation. For example, the second 3D rendering unit 630 may
perform a 3D rendering operation using an HRTF having different coefficients from

WO 2007/091842 PCT/KR2007/000668
those of an HRTF used by an encoding apparatus.
[86] The filter information storage unit 640 stores filter information regarding a filter
used to perform a 3D rendering, for example, HRTF coefficient information. The
second 3D rendering unit 630 may generate a combined 3D down-mix using the filter
information stored in the filter information storage unit 640.
[87] The filter information storage unit 640 may store a plurality of pieces of filter in-
formation respectively corresponding to a plurality of filters. In this case, one of the
plurality of pieces of filter information may be selected at a user's choice or according
to the capabilities of the decoding apparatus or required sound quality.
[88] People from different races may have different ear structures. Thus, HRTF co-
efficients optimized for different individuals may differ from one another. The
decoding apparatus illustrated in FIG. 6 can generate a 3D down-mix signal optimized
for the user. In addition, the decoding apparatus illustrated in FIG. 6 can generate a 3D
down-mix signal with 3D effects corresponding to an HRTF filter desired by the user,
regardless of the type of HRTF provided by a 3D down-mix signal provider.
[89] FIG. 7 is a block diagram of a 3D rendering apparatus according to an embodiment
of the present invention. Referring to FIG. 7, the 3D rendering apparatus includes first
and second domain conversion units 700 and 720 and a 3D rendering unit 710. In order
to perform a 3D rendering operation in a predetermined domain, the first and second
domain conversion units 700 and 720 may be respectively disposed in front of and
behind the 3D rendering unit 710.
[90] Referring to FIG. 7, an input down-mix signal is converted into a frequency-
domain down-mix signal by the first domain conversion unit 700. More specifically,
the first: domain conversion unit 700 may convert the input down-mix signal into a
DFT-domain down-mix signal or a FFT-domain down-mix signal by performing DFT
orFFT.
[91] The 3D rendering unit 710 generates a multi-channel signal by applying spatial in-
formation to the frequency-domain down-mix signal provided by the first domain
conversion unit 700. Thereafter, the 3D rendering unit 710 generates a 3D down-mix
signal by filtering the multi-channel signal.
[92] The 3D down-mix signal generated by the 3D rendering unit 710 is converted into
a time-domain 3D down-mix signal by the second domain conversion unit 720. More
specifically, the second domain conversion unit 720 may perform IDFT or IFFT on the
3D down-mix signal generated by the 3D rendering unit 710.
[93] During the conversion of a frequency-domain 3D down-mix signal into a time-
domain 3D down-mix signal, data loss or data distortion such as aliasing may occur.
[94] In order to generate a multi-channel signal and a 3D down-mix signal in a
frequency domain, spatial information for each parameter band may be mapped to the

WO 2007/091842 PCT/KR2007/000668
frequency domain, and a number of filter coefficients may be converted to the
frequency domain.
[95] The 3D rendering unit 710 may generate a 3D down-mix signal by multiplying the
frequency-domain down-mix signal provided by the first domain conversion unit 700,
the spatial information, and the filter coefficients.
[96] A time-domain signal obtained by multiplying a down-mix signal, spatial in-
formation and a plurality of filter coefficients that are all represented in an M-point
frequency domain has M valid signals. In order to represent the down-mix signal, the
spatial information and the filter in the M-point frequency domain, M-point DFT or M-
point FFT may be performed.
[97] Valid signals are signals that do not necessarily have a value of 0. For example, a
total of x valid signals can be generated by obtaining x signals from an audio signal
through sampling. Of the x valid signals, y valid signals may be zero-padded. Then, the
number of valid signals is reduced to (x-y). Thereafter, a signal with a valid signals and
a signal with b valid signals are convoluted, thereby obtaining a total of (a+b-1) valid
signals,
[98] The multiplication of the down-mix signal, the spatial information, and the filter
coefficients in the M-point frequency domain can provide the same effect as
convoluting the down-mix signal, the spatial information, and the filter coefficients in
a time-domain. A signal with (3*M-2) valid signals can be generated by converting the
down-mix signal, the spatial information and the filter coefficients in the M-point
frequency domain to a time domain and convoluting the results of the conversion.
[99] Therefore, the number of valid signals of a signal obtained by multiplying a down-
mix signal, spatial information, and filter coefficients in a frequency domain and
converting the result of the multiplication to a time domain may differ from the
number of valid signals of a signal obtained by convoluting the down-mix signal, the
spatial information, and the filter coefficients in the time domain. As a result, aliasing
may occur during the conversion of a 3D down-mix signal in a frequency domain into
a time-domain signal.
[100] In order to prevent aliasing, the sum of the number of valid signals of a down-mix
signal in a time domain, the number of valid signals of spatial information mapped to a
frequency domain, and the number of filter coefficients must not be greater than M.
The number of valid signals of spatial information mapped to a frequency domain may
be determined by the number of points of the frequency domain. In other words, if
spatial information represented for each parameter band is mapped to an N-point
frequency domain, the number of valid signals of the spatial information may be N.
[101] Referring to FIG. 7, the first domain conversion unit 700 includes a first zero-
padding unit 701 and a first frequency-domain conversion unit 702. The third

14
WO 2007/091842 PCT/KR2007/000668
rendering unit 710 includes a mapping unit 711, a time-domain conversion unit 712, a
second zero-padding unit 713, a second frequency-domain conversion unit 714, a
multi-channel signal generation unit 715, a third zero-padding unit 716, a third
frequency-domain conversion unit 717, and a 3D down-mix signal generation unit 718.
[102] The first zero-padding unit 701 performs a zero-padding operation on a down-mix
signal with X samples in a time domain so that the number of samples of the down-mix
signal can be increased from X to M. The first frequency-domain conversion unit 702
converts the zero-padded down-mix signal into an M-point frequency-domain signal.
The zero-padded down-mix signal has M samples. Qf the M samples of the zero-
padded down-mix signal, only X samples are valid signals.
[103] The mapping unit 711 maps spatial information for each parameter band to an N-
point frequency domain. The time-domain conversion unit 712 converts spatial in-
formation obtained by the mapping performed by the mapping unit 711 to a time
domain. Spatial information obtained by the conversion performed by the time-domain
conversion unit 712 has N samples.
[104] The second zero-padding unit 713 performs a zero-padding operation on the spatial
information with N samples in the time domain so that the number of samples of the
spatial information can be increased from N to M. The second frequency-domain
conversion unit 714 converts the zero-padded spatial information into an M-point
frequency-domain signal. The zero-padded spatial information has N samples. Of the
N samples of the zero-padded spatial information, only N samples are valid.
[105] The multi-channel signal generation unit 715 generates a multi-channel signal by
multiplying the down-mix signal provided by the first frequency-domain conversion
unit 712 and spatial information provided by the second frequency-domain conversion
unit 714. The multi-channel signal generated by the multi-channel signal generation
unit 715 has M valid signals. On the other hand, a multi-channel signal obtained by
convoluting, in the time domain, the down-mix signal provided by the first frequency-
domain conversion unit 712 and the spatial information provided by the second
frequency-domain conversion unit 714 has (X+N-l) valid signals.
[106] The third zero-padding unit 716 may perform a zero-padding operation on Y filter
coefficients that are represented in the time domain so that the number of samples can
be increased to M. The third frequency-domain conversion unit 717 converts the zero-
padded filter coefficients to the M-point frequency domain. The zero-padded filter co-
efficients have M samples. Of the M samples, only Y samples are valid signals.
[107] The 3D down-mix signal generation unit 718 generates a 3D down-mix signal by
multiplying the multi-channel signal generated by the multi-channel signal generation
unit 715 and a plurality of filter coefficients provided by the third frequency-domain
conversion unit 717. The 3D down-mix signal generated by the 3D down-mix signal

15
WO 2007/091842 PCT/KR2007/000668
generation unit 718 has M valid signals. On the other hand, a 3D down-mix signal
obtained by convoluting, in the time domain, the multi-channel signal generated by the
multi-channel signal generation unit 715 and the filter coefficients provided by the
third frequency-domain conversion unit 717 has (X+N+Y-2) valid signals.
[108] It is possible to prevent aliasing by setting the M-point frequency domain used by
the first, second, and third frequency-domain conversion units 702, 714, and 717 to
satisfy the following equation: M>(X+N+Y-2). In other words, it is possible to prevent
aliasing by enabling the first, second, and third frequency-domain conversion units
702, 714, and 717 to perform M-point DFT or M-point FFT that satisfies the following
equation: M>(X+N+Y-2).
[109] The conversion to a frequency domain may be performed using a filter bank other
than a DFT filter bank, an FFT filter bank, and QMF bank. The generation of a 3D
down-mix signal may be performed using an HRTF filter.
[110] The number of valid signals of spatial information may be adjusted using a method
other than the above-mentioned methods or may be adjusted using one of the above-
mentioned methods that is most efficient and requires the least amount of computation.
[Ill] Aliasing may occur not only during the conversion of a signal, a coefficient or
spatial information from a frequency domain to a time domain or vice versa but also
during the conversion of a signal, a coefficient or spatial information from a QMF
domain to a hybrid domain or vice versa. The above-mentioned methods of preventing
aliasing may also be used to prevent aliasing from occurring during the conversion of a
signal, a coefficient or spatial information from a QMF domain to a hybrid domain or
vice versa.
[112] Spatial information used to generate a multi-channel signal or a 3D down-mix
signal may vary. As a result of the variation of the spatial information, signal discon-
tinuities may occur as noise in an output signal.
[113] Noise in an output signal may be reduced using a smoothing method by which
spatial information can be prevented from rapidly varying.
[114] For example, when first spatial information applied to a first frame differs from
second spatial information applied to a second frame when the first frame and the
second frame are adjacent to each other, a discontinuity is highly likely to occur
between the first and second frames.
[115] In this case, the second spatial information may be compensated for using the first
spatial information or the first spatial information may be compensated for using the
second spatial information so that the difference between the first spatial information
and the second spatial information can be reduced, and that noise caused by the dis-
continuity between the first and second frames can be reduced. More specifically, at
least one of the first spatial information and the second spatial information may be

16
WO 2007/091842 PC17KR2007/000668
replaced with the average of the first spatial information and the second spatial in-
formation, thereby reducing noise.
[116] Noise is also likely to be generated due to a discontinuity between a pair of
adjacent parameter bands. For example, when third spatial information corresponding
to a first parameter band differs from fourth spatial information corresponding to a
second parameter band when the first and second parameter bands are adjacent to each
other, a discontinuity is likely to occur between the first and second parameter bands.
[117] In this case, the third spatial information may be compensated for using the fourth
spatial information or the fourth spatial information may be compensated for using the
third spatial information so that the difference between the third spatial information
and the fourth spatial information can be reduced, and that noise caused by the dis-
continuity between the first and second parameter bands can be reduced. More
specifically, at least one of the third spatial information and the fourth spatial in-
formation may be replaced with the average of the third spatial information and the
fourth spatial information, thereby reducing noise.
[118] Noise caused by a discontinuity between a pair of adjacent frames or a pair of
adjacent parameter bands may be reduced using methods other than the above-
mentioned methods.
[119] More specifically, each frame may be multiplied by a window such as a Hanning
window, and an "overlap and add" scheme may be applied to the results of the multi-
plication so that the variations between the frames can be reduced. Alternatively, an
output signal to which a plurality of pieces of spatial information are applied may be
smoothed so that variations between a plurality of frames of the output signal can be
prevented.
[120] The decorrelation between channels in a DFT domain using spatial information, for
example, ICC, may be adjusted as follows.
[121] The degree of decorrelation may be adjusted by multiplying a coefficient of a
signal input to a one-to-two (OTT) or two-to-three (TTT) box by a predetermined
value. The predetermined value can be defined by the following equation:
(A+(l-A*A)A0.5*i) where A indicates an ICC value applied to a predetermined band
of the OTT or TTT box and i indicates an imaginary part. The imaginary part may be
positive or negative.
[122] The predetermined value may accompany a weighting factor according to the char-
acteristics of the signal, for example, the energy level of the signal, the energy charac-
teristics of each frequency of the signal, or the type of box to which the ICC value A is
applied. As a result of the introduction of the weighting factor, the degree of
decorrelation may be further adjusted, and interframe smoothing or interpolation may
be applied.

17
WO 2007/091842 PCT/KR2007/000668
[123] As described above with reference to FIG. 7, a 3D down-mix signal may be
generated in a frequency domain by using an HRTF or a head related impulse response
(HRIR), which is converted to the frequency domain..
[124] Alternatively, a 3D down-mix signal may be generated by convoluting an HRIR
and a down-mix signal in a time domain. A 3D down-mix signal generated in a
frequency domain may be left in the frequency domain without being subjected to
inverse domain transform.
[125] In order to convolute an HRIR and a down-mix signal in a time domain, a finite
impulse response (FIR) filter or an infinite impulse response (IIR) filter may be used.
[126] As described above, an encoding apparatus or a decoding apparatus according to an
embodiment of the present invention may generate a 3D down-mix signal using a first
method that involves the use of an HRTF in a frequency domain or an HRIR converted
to the frequency domain, a second method that involves convoluting an HRIR in a time
domain, or the combination of the first and second methods.
[127] FIGS. 8 through 11 illustrate bitstreams according to embodiments of the present
invention.
[128] Referring to FIG. 8, a bitstream includes a multi-channel decoding information
field which includes information necessary for generating a multi-channel signal, a 3D
rendering information field which includes information necessary for generating a 3D
down-mix signal, and a header field which includes header information necessary for
using the information included in the multi-channel decoding information field and the
information included in the 3D rendering information field. The bitstream may include
only one or two of the multi-channel decoding information field, the 3D rendering in-
formation field, and the header field.
[129] Referring to FIG. 9, a bitstream, which contains side information necessary for a
decoding operation, may include a specific configuration header field which includes
header information of a whole encoded signal and a plurality of frame data fields
which includes side information regarding a plurality of frames. More specifically,
each of the frame data fields may include a frame header field which includes header
information of a corresponding frame and a frame parameter data field which includes
spatial information of the corresponding frame. Alternatively, each of the frame data
fields may include a frame parameter data field only.
[130] Each of the frame parameter data fields may include a plurality of modules, each
module including a flag and parameter data. The modules are data sets including
parameter data such as spatial information and other data such as down-mix gain and
smoothing data which is necessary for improving the sound quality of a signal.
[131] If module data regarding information specified by the frame header fields is
received without any additional flag, if the information specified by the frame header

18
WO 2007/091842 PCT7KR2007/000668
fields is further classified, or if an additional flag and data are received in connection
with information not specified by the frame header, module data may not include any
flag.
[132] Side information regarding a 3D down-mix signal, for example, HRTF coefficient
information, may be included in at least one of the specific configuration header field,
the frame header fields, and the frame parameter data fields.
[133] Referring to FIG. 10, a bitstream may include a plurality of multi-channel decoding
information fields which include information necessary for generating multi-channel
signals and a plurality of 3D rendering information fields which include information
necessary for generating 3D down-mix signals.
[134] When receiving the bitstream, a decoding apparatus may use either the multi-
channel decoding information fields or the 3D rendering information field to perform a
decoding operation and skip whichever of the multi-channel decoding information
fields and the 3D rendering information fields are not used in the decoding operation.
In this case, it may be determined which of the multi-channel decoding information
fields and the 3D rendering information fields are to be used to perform a decoding
operation according to the type of signals to be reproduced.
[135] In other words, in order to generate multi-channel signals, a decoding apparatus
. may skip the 3D rendering information fields, and read information included in the
multi-channel decoding information fields. On the other hand, in order to generate 3D
down-mix signals, a decoding apparatus may skip the multi-channel decoding in-
formation fields, and read information included in the 3D rendering information fields.
[136] Methods of skipping some of a plurality of fields in a bitstream are as follows.
[137] First, field length information regarding the size in bits of a field may be included
in a bitstream. In this case, the field may be skipped by skipping a number of bits cor-
responding to the size in bits of the field. The field length information may be disposed
at the beginning of the field.
[138] Second, a syncword may be disposed at the end or the beginning of a field. In this
case, the field may be skipped by locating the field based on the location of the
syncword.
[139] Third, if the length of a field is determined in advance and fixed, the field may be
skipped by skipping an amount of data corresponding to the length of the field. Fixed
field length information regarding the length of the field may be included in a
bitstream or may be stored in a decoding apparatus.
[140] Fourth, one of a plurality of fields may be skipped using the combination of two or
more of the above-mentioned field skipping methods.
[141] Field skip information, which is information necessary for skipping a field such as
field length information, syncwords, or fixed field length information may be included

WO 2007/091842 PCT/KR2007/000668
in one of the specific configuration header field, the frame header fields, and the frame
parameter data fields illustrated in FIG. 9 or may be included in a field other than those
illustrated in FIG. 9.
[142] For example, in order to generate multi-channel signals, a decoding apparatus may
skip the 3D rendering information fields with reference to field length information, a
syncword, or fixed field length information disposed at the beginning of each of the 3D
rendering information fields, and read information included in the multi-channel
decoding information fields.
[143] On the other hand, in order to generate 3D down-mix signals, a decoding apparatus
may skip the multi-channel decoding information fields with reference to field length
information, a syncword, or fixed field length information disposed at the beginning of
each of the multi-channel decoding information fields, and read information included
in the 3D rendering information fields.
[144] A bitstream may include information indicating whether data included in the
bitstream is necessary for generating multi-channel signals or for generating 3D down-
mix signals.
[145] However, even if a bitstream does not include any spatial information such as CLD
but includes only data (e.g., HRTF filter coefficients) necessary for generating a 3D
down-mix signal, a multi-channel signal can be reproduced through decoding using the
data necessary for generating a 3D down-mix signal without a requirement of the
spatial information.
[146] For example, a stereo parameter, which is spatial information regarding two
channels, is obtained from a down-mix signal. Then, the stereo parameter is converted
into spatial information regarding a plurality of channels to be reproduced, and a multi-
channel signal is generated by applying the spatial information obtained by the
conversion to the down-mix signal.
[147] On the other hand, even if a bitstream includes only data necessary for generating a
multi-channel signal, a down-mix signal can be reproduced without a requirement of
an additional decoding operation or a 3D down-mix signal can be reproduced by
performing 3D processing on the down-mix signal using an additional HRTF filter.
[148] If a bitstream includes both data necessary for generating a multi-channel signal
and data necessary for generating a 3D down-mix signal, a user may be allowed to
decide whether to reproduce a multi-channel signal or a 3D down-mix signal.
[ 149] Methods of skipping data will hereinafter be described in detail with reference to
respective corresponding syntaxes.
[150] Syntax 1 indicates a method of decoding an audio signal in units of frames.
[151]
[152] [Syntax 1]

WO 2007/091842

PCT7KR2007/000668



[153]
[154]
[155]
[156]
[157]
[158]
[159]
[160]


In Syntax 1, Ottdata() and TttData() are modules which represent parameters (such
as spatial information including a CLD, ICC, and CPC) necessary for restoring a multi-
channel signal from a down-mix signal, and SmgData(), TempShapeData(), Arbitrary-
DownmixData(), and ResidualData() are modules which represent information
necessary for improving the quality of sound by correcting signal distortions that may
have occurred during an encoding operation.
For example, if a parameter such as a CLD, ICC or CPC and information included
in the module ArbitraryDownmixData() are only used during a decoding operation, the
modules SmgData() and TempShapeData(), which are disposed between the modules
TttData() and ArbitraryDownmixData(), may be unnecessary. Thus, it is efficient to
skip the modules SmgData() and TempShapeData().
A method of skipping modules according to an embodiment of the present
invention will hereinafter be described in detail with reference to Syntax 2 below.
[Syntax 2]

WO 2007/091842 PCT/KR2007/000668

[161]
[162] Referring to Syntax 2, a module SkipData() may be disposed in front of a module
to be skipped, and the size in bits of the module to be skipped is specified in the
module SkipData() as bsSkipBits.
[163] In other words, assuming that modules SmgDataQ and TempShapeData() are to be
skipped, and that the size in bits of the modules SmgData() and TempShapeData()
combined is 150, the modules SmgData() and TempShapeData() can be skipped by
setting bsSkipBits to 150.
[164] A method of skipping modules according to another embodiment of the present
invention will hereinafter be described in detail with reference to Syntax 3.
[165]
[166] [Syntax 3]
[167]
[168]
[169] Referring to Syntax 3, an unnecessary module may be skipped by using
bsSkipSyncflag, which is a flag indicating whether to use a syncword, and

WO 2007/091842 PCT7KR2007/000668
bsSkipSyncword, which is a syncword that can be disposed at the end of a module to
be skipped.
[170] More specifically, if the flag bsSkipSyncflag is set such that a syncword can be
used, one or more modules between the flag bsSkipSyncflag and the syncword
bsSkipSyncword, i.e., modules SmgData() and TempShapeData(), may be skipped.
[171] Referring to FIG. 11, a bitstream may include a multi-channel header field which
includes header information necessary for reproducing a multi-channel signal, a 3D
rendering header field which includes header information necessary for reproducing a
3D down-mix signal, and a plurality of multi-channel decoding information fields,
which include data necessary for reproducing a multi- channel signal.
[172] In order to reproduce a multi-channel signal, a decoding apparatus may skip the 3D
rendering header field, and read data from the multi-channel header field and the multi-
channel decoding information fields.
[173] A method of skipping the 3D rendering header field is the same as the field
skipping methods described above with reference to FTC. 10, and thus, a detailed de-
scription thereof will be skipped.
[174] In order to reproduce a 3D down-mix signal, a decoding apparatus may read data
from the multi-channel decoding information fields and the 3D rendering header field.
For example, a decoding apparatus may generate a 3D down-mix signal using a down-
mix signal included in the multi-channel decoding information field and HRTF co-
efficient information included in the 3D down-mix signal.
[175] FIG. 12 is a block diagram of an encoding/decoding apparatus for processing an
arbitrary down-mix signal according to an embodiment of the present invention.
Referring to FIG. 12, an arbitrary down-mix signal is a down-mix signal other than a
down-mix signal generated by a multi-channel encoder 801 included in an encoding
apparatus 800. Detailed descriptions of the same processes as those of the embodiment
of FIG. 1 will be omitted.
[176] Referring to FIG. 12, the encoding apparatus 800 includes the multi-channel
encoder 801, a spatial information synthesization unit 802, and a comparison unit 803.
[177] The multi-channel encoder 801 down-mixes an input multi-channel signal into a
stereo or mono down-mix signal, and generates basic spatial information necessary for
restoring a multi-channel signal from the down-mix signal.
[178] The comparison unit 803 compares the down-mix signal with an arbitrary down-
mix signal, and generates compensation information based on the result of the
comparison. The compensation information is necessary for compensating for the
arbitrary down-mix signal so that the arbitrary down-mix signal can be converted to be
approximate to the down-mix signal. A decoding apparatus may compensate for the
arbitrary down-mix signal using the compensation information and restore a multi-

WO 2007/091842 PCT/KR2007/000668
channel signal using the compensated arbitrary down-mix signal. The restored multi-
channel signal is more similar than a multi-channel signal restored from the arbitrary
down-mix signal generated by the multi-channel encoder 801 to the original input
multi-channel signal.
[179] The compensation information may be a difference between the down-mix signal
and the arbitrary down-mix signal. A decoding apparatus may compensate for the
arbitrary down-mix signal by adding, to the arbitrary down-mix signal, the difference
between the down-mix signal and the arbitrary down-mix signal.
[180] The. difference between the down-mix signal and the arbitrary down-mix signal
may be down-mix gain which indicates the difference; between the energy levels of the
down-mix signal and the arbitrary down-mix signal.
[181] The down-mix gain may be determined for each frequency band, for each time/
time slot, and/or for each channel. For example, one part of the down-mix gain may be
determined for each frequency band, and another part of the down-mix gain may be
determined for each time slot.
[182] The down-mix gain may be determined for each parameter band or for each
frequency band optimized for the arbitrary down-mix signal. Parameter bands are
frequency intervals to which parameter-type spatial information is applied.
[183] The difference between the energy levels of the down-mix signal and the arbitrary
down-mix signal may be quantized. The resolution of quantization levels for
quantizing the difference between the energy levels of the down-mix signal and the
arbitrary down-mix signal may be the same as or different from the resolution of
quantization levels for quantizing a CLD between the down-mix signal and the
arbitrary down-mix signal. In addition, the quantization of the difference between the
energy levels of the down-mix signal and the arbitrary down-mix signal may involve
the use of all or some of the quantization levels for quantizing the CLD between the
down-mix signal and the arbitrary down-mix signal.
[184] Since the resolution of the difference between the energy levels of the down-mix
signal and the arbitrary down-mix signal is generally lower than the resolution of the
CLD between the down-mix signal and the arbitrary down-mix signal, the resolution of
the quantization levels for quantizing the difference between the energy levels of the
down-mix signal and the arbitrary down-mix signal may have a minute value
compared to the resolution of the quantization levels for quantizing the CLD between
the down-mix signal and the arbitrary down-mix signal.
[185] The compensation information for compensating for the arbitrary down-mix signal
may be extension information including residual information which specifies
components of the input multi-channel signal that cannot be restored using the
arbitrary down-mix signal or the down-mix gain. A decoding apparatus can restore

WO 2007/091842 PCT/KR2007/000668
components of the input multi-channel signal that cannot be restored using the
arbitrary down-mix signal or the down-mix gain using the extension information,
thereby restoring a signal almost indistinguishable from the original input multi-
channel signal.
[186] Methods of generating the extension information are as follows.
[187] The multi-channel encoder 801 may generate information regarding components of
the input multi-channel signal that are lacked by the down-mix signal as first extension
information. A decoding apparatus may restore a signal almost indistinguishable from
the original input multi-channel signal by applying the first extension information to
the generation of a multi-channel signal using the down-mix signal and the basic
spatial information.
[188] Alternatively, the multi-channel encoder 801 may restore a multi-channel signal
using the down-mix signal and the basic spatial information, and generate the
difference between the restored multi-channel signal and the original input multi-
channel signal as the first extension information.
[189] The comparison unit 803 may generate, as second extension information, in-
formation regarding components of the down-mix signal that are lacked by the
arbitrary down-mix signal, i.e., components of the down-mix signal that cannot be
compensated for using the down-mix gain. A decoding apparatus may restore a signal
almost indistinguishable from the down-mix signal using the arbitrary down-mix signal
and the second extension information.
[190] The extension information may be generated using various residual coding
methods other than the above-described method.
[191] The down-mix gain and the extension information may both be used as com-
pensation information. More specifically, the down-mix gain and the extension in-
formation may both be obtained for an entire frequency band of the down-mix signal
and may be used together as compensation information. Alternatively, the down-mix
gain may be used as compensation information for one part of the frequency band of
the down-mix signal, and the extension information may be used as compensation in-
formation for another part of the frequency band of the down-mix signal. For example,
the extension information may be used as compensation information for a low
frequency band of the down-mix signal, and the down-mix gain may be used as comp
ensation information for a high frequency band of the down-mix signal.
[ 192] Extension information regarding portions of the down-mix signal, other than the
low-frequency band of the down-mix signal, such as peaks or notches that may con-
siderably affect the quality of sound may also be used as compensation information.
[193] The spatial information synthesization unit 802 synthesizes the basic spatial in-
formation (e.g., a CLD, CPC, ICC, and CTD) and the compensation information,

WO 2007/091842 PCT/KR2007/000668
thereby generating spatial information. In other words, the spatial information, which
is transmitted to a decoding apparatus, may include the basic spatial information, the
down-mix gain, and the first and second extension information.
[194] The spatial information may be included in a bitstream along with the arbitrary
down-mix signal, and the bitstream may be transmitted to a decoding apparatus.
[195] The extension information and the arbitrary down-mix signal may be encoded
using an audio encoding method such as an AAC method, a MP3 method, or a BSAC
method. The extension information and the arbitrary down-mix signal may be encoded
using the same audio encoding method or different audio encoding methods.
[196] If the extension information and the arbitrary down-mix signal are encoded using
the same audio encoding method, a decoding apparatus may decode both the extension
information and the arbitrary down-mix signal using a single audio decoding method.
In this case, since the arbitrary down-mix signal can always be decoded, the extension
information can also always be decoded. However, since the arbitrary down-mix signal
is generally input to a decoding apparatus as a pulse code modulation (PCM) signal,
the type of audio codec used to encode the arbitrary down-mix signal may not be
readily identified, and thus, the type of audio codec used to encode the extension in-
formation may not also be readily identified.
[197] Therefore, audio codec information regarding the type of audio codec used to
encode the arbitrary down-mix signal and the extension information may be inserted
into a bitstream.
[198] More specifically, the audio codec information may be inserted into a specific con-
figuration header field of a bitstream. In this case, a decoding apparatus may extract
the audio codec information from the specific configuration header field of the
bitstream and use the extracted audio codec information to decode the arbitrary d own-
mix signal and the extension information.
[199] On the other hand, if the arbitrary down-mix signal and the extension information
are encoded using different audio encoding methods, the extension information may
not be able to be decoded. In this case, since the end of the extension information
cannot be identified, no further decoding operation can be performed.
[200] In order to address this problem, audio codec information regarding the types of
audio codecs respectively used to encode the arbitrary down-mix signal and the
extension information may be inserted into a specific configuration header field of a
bitstream. Then, a decoding apparatus may read the audio codec information from the
specific configuration header field of the bitstream and use the read information to
decode the extension information. If the decoding apparatus does not include any
decoding unit that can decode the extension information, the decoding of the extension
information may not further proceed, and information next to the extension in-

26
WO 2007/091842 PCT/KR2007/000668
formation may be read.
[201] Audio codec information regarding the type of audio codec used to encode the
extension information may be represented by a syntax element included in a specific
configuration header field of a bitstream. For example, the audio codec information
may be represented by bsResidualCodecType, which is a 4-bit syntax element, as
indicated in Table 1 below.
[202]
[203] Table 1
[204]
bsResidualCodecType Codec
0 AAC
1 MP3
2 BSAC
3... 15 Reserved
[205]
[206] The extension information may include not only the residual information but also
channel expansion information. The channel expansion information is information
necessary for expanding a multi-channel signal obtained through decoding using the
spatial information into a multi-channel signal with more channels. For example, the
channel expansion information may be information necessary for expanding a
5.1-channel signal or a 7.1-channel signal into a 9.1-channel signal.
[207] The extension information may be included in a bitstream, and the bitstream may
be transmitted to a decoding apparatus. Then, the decoding apparatus may compensate
for the down-mix signal or expand a multi-channel signal using the extension in-
formation. However, the decoding apparatus may skip the extension information,
instead of extracting the extension information from the bitstream. For example, in the
case of generating a multi-channel signal using a 3D down-mix signal included in the
bitstream or generating a 3D down-mix signal using a down-mix signal included in the
bitstream, the decoding apparatus may skip the extension information.
[208] A method of skipping the extension information included in a bitstream may be the
same as one of the field skipping methods described above with reference to FIG. 10.
[209] For example, the extension information may be skipped using at least one of bit
size information which is attached to the beginning of a bitstream including the
extension information and indicates the size in bits of the extension information, a
syncword which is attached to the beginning or the end of the field including the

WO 2007/091842 PCT/KR2007/000668
extension information, and fixed bit size information which indicates a fixed size in
bits of the extension information. The bit size information, the syncword, and the fixed
bit size information may all be included in a bitstream. The fixed bit size information
may also be stored in a decoding apparatus.
[210] Referring to FIG. 12, a decoding unit 810 includes a down-mix compensation unit
811, a 3D rendering unit 815, and a multi-channel decoder 816.
[211] The down-mix compensation unit 811 compensates for an arbitrary down-mix
signal using compensation information included in spatial information, for example,
using down-mix gain or extension information.
[212] The 3D rendering unit 815 generates a decoder 3D down-mix signal by performing
a 3D rendering operation on the compensated down-mix signal. The multi-channel
decoder 816 generates a 3D multi-channel signal using the compensated down-mix
signal and basic spatial information, which is included in the spatial information.
[213] The down-mix compensation unit 811 may compensate for the arbitrary down-mix
signal in the following manner.
[214] If the compensation information is down-mix gain, the down-mix compensation
unit 811 compensates for the energy level of the arbitrary down-mix signal using the
down-mix gain so that the arbitrary down-mix signal can be converted into a signal
similar to a down-mix signal.
[215] If the compensation information is second extension information, the down-mix
compensation unit 811 may compensate for components that are lacked by the
arbitrary down-mix signal using the second extension information.
[216] The multi-channel decoder 816 may generate a multi-channel signal by se-
quentially applying pre-matrix Ml, mix-matrix M2 and post-matrix M3 to a down-mix
signal. In this case, the second extension information may be used to compensate for
the down-mix signal during the application of mix-matrix M2 to the down-mix signal.
In other words, the second extension information may be used to compensate for a
down-mix signal to which pre-matrix Ml has already been applied.
[217] As described above, each of a plurality of channels may be selectively
compensated for by applying the extension information to the generation of a multi-
channel signal. For example, if the extension information is applied to a center channel
of mix-matrix M2, left- and right-channel components of the down-mix signal may be
compensated for by the extension information. If the extension information is applied
to a left channel of mix-matrix M2, the left-channel component of the down-mix signal
may be compensated for by the extension information.
[218] The down-mix gain and the extension information may both be used as the com-
pensation information. For example, a low frequency band of the arbitrary down-mix
signal may be compensated for using the extension information, and a high frequency

WO 2007/091842 PCT/KR2007/000668
band of the arbitrary down-mix signal may be compensated for using the down-mix
gain. In addition, portions of the arbitrary down-mix signal, other than the low
frequency band of the arbitrary down-mix signal, for example, peaks or notches that
may considerably affect the quality of sound, may also be compensated for using the
extension information. Information regarding portion to be compensated for by the
extension information may be included in a bitstream. Information indicating whether
a down-mix signal included in a bitstream is an arbitrary down-mix signal or not and
information indicating whether the bitstream includes compensation information may
be included in the bitstream.
[219] In order to prevent clipping of a down-mix signal generated by the encoding unit
800, the down-mix signal may be divided by predetermined gain. The predetermined
gain may have a static value or a dynamic value.
[220] The down-mix compensation unit 811 may restore the original down-mix signal by
compensating for the down-mix signal, which is weakened in order to prevent clipping,
using the predetermined gain.
[221] An arbitrary down-mix signal compensated for by the down-mix compensation unit
811 can be readily reproduced. Alternatively, an arbitrary down-mix signal yet to be
compensated for may be input to the 3D rendering unit 815, and may be converted into
a decoder 3D down-mix signal by the 3D rendering unit 815.
[222] Referring to FIG. 12, the down-mix compensation unit 811 includes a first domain
converter 812, a compensation processor 813, and a second domain converter 814.
[223] The first domain converter 812 converts the domain of an arbitrary down-mix
signal into a predetermined domain. The compensation processor 813 compensates for
the arbitrary down-mix signal in the predetermined domain, using compensation in-
formation, for example, down-mix gain or extension information.
[224] The compensation of the arbitrary down-mix signal may be performed in a QMF/
hybrid domain. For this, the first domain converter 812 may perform QMF/hybrid
analysis on the arbitrary down-mix signal. The first domain converter 812 may convert
the domain of the arbitrary down-mix signal into a domain, other than a QMF/hybrid
domain, for example, a frequency domain such as a DFT or FFT domain. The com-
pensation of the arbitrary down-mix signal may also be performed in a domain, other
than a QMF/hybrid domain, for example, a frequency domain or a time domain.
[225] The second domain converter 814 converts the domain of the compensated
arbitrary down-mix signal into the same domain as the original arbitrary down-mix
signal. More specifically, the second domain converter 814 converts the domain of the
compensated arbitrary down-mix signal into the same domain as the original arbitrary
down-mix signal by inversely performing a domain conversion operation performed by
the first domain converter 812.

WO 2007/091842 PCT/KR2007/000668
[226] For example, the second domain converter 814 may convert the compensated
arbitraiy down-mix signal into a time-domain signal by performing QMF/hybrid
synthesis on the compensated arbitrary down-mix signal. Also, the second domain
converter 814 may perform IDFT or IFFT on the compensated arbitrary down-mix
signal.
[227] The 3D rendering unit 815, like the 3D rendering unit 710 illustrated in FIG. 7,
may perform a 3D rendering operation on the compensated arbitrary down-mix signal
in a frequency domain, a QMF/hybrid domain or a time domain. For this, the 3D
rendering unit 815 may include a domain converter (not shown). The domain converter
converts the domain of the compensated arbitrary down-mix signal into a domain in
which a 3D rendering operation is to be performed or converts the domain of a signal
obtained by the 3D rendering operation.
[228] The domain in which the compensation processor 813 compensates for the
arbitrary down-mix signal may be the same as or different from the domain in which
the 3D rendering unit 815 performs a 3D rendering operation on the compensated
arbitrary down-mix signal.
[229] FIG. 13 is a block diagram of a down-mix compensation/3D rendering unit 820
according to an embodiment of the present invention. Referring to FIG. 13, the down-
mix compensation/3D rendering unit 820 includes a first domain converter 821, a
second domain converter 822, a compensation/3D rendering processor 823, and a third
domain converter 824.
[230] The down-mix compensation/3D rendering unit 820 may perform both a com-
pensation operation and a 3D rendering operation on an arbitrary down-mix signal in a
single domain, thereby reducing the amount of computation of a decoding apparatus.
[231] More specifically, the first domain converter 821 converts the domain of the
arbitrary down-mix signal into a first domain in which a compensation operation and a
3D rendering operation are to be performed. The second domain converter 822
converts spatial information, including basic spatial information necessary for
generating a multi-channel signal and compensation information necessary for com-
pensating for the arbitrary down-mix signal, so that the spatial information can become
applicable in the first domain. The compensation information may include at least one
of down-mix gain and extension information.
[232] For example, the second domain converter 822 may map compensation in-
formation corresponding to a parameter band in a QMF/hybrid domain to a frequency
band so that the compensation information can become readily applicable in a
frequency domain.
[233] The first domain may be a frequency domain such as a DFT or FFT domain, a
QMF/hybrid domain, or a time domain. Alternatively, the first domain may be a

WO 2007/091842 PCT7KR2007/000668
domain other than those set forth herein.
[234] During the conversion of the compensation information, a time delay may occur. In
order to address this problem, the second domain converter 822 may perform a time
delay compensation operation so that a time delay between the domain of the com-
pensation information and the first domain can be compensated for.
[235] The compensation/3D rendering processor 823 performs a compensation operation
on the arbitrary down-mix signal in the first domain using the converted spatial in-
formation and then performs a 3D rendering operation on a signal obtained by the
compensation operation. The compensation/3D rendering processor 823 may perform a
compensation operation and a 3D rendering operation in a different order from that set
forth herein.
[236] The compensation/3D rendering processor 823 may perform a compensation
operation and a 3D rendering operation on the arbitrary down-mix signal at the same
time. For example, the compensation/3D rendering processor 823 may generate a
compensated 3D down-mix signal by performing a 3D rendering operation on the
arbitrary down-mix signal in the first domain using a new filter coefficient, which is
the combination of the compensation information and an existing filter coefficient
typically used in a 3D rendering operation.
[237] The third domain converter 824 converts the domain of the 3D down-mix signal
generated by the compensation/3D rendering processor 823 into a frequency domain.
[238] FIG. 14 is a block diagram of a decoding apparatus 900 for processing a
compatible down-mix signal according to an embodiment of the present invention.
Referring to FIG. 14, the decoding apparatus 900 includes a first multi-channel
decoder 910, a down-mix compatibility processing unit 920, a second multi-channel
decoder 930, and a 3D rendering unit 940. Detailed descriptions of the same decoding
processes as those of the embodiment of FIG. 1 will be omitted.
[239] A compatible down-mix signal is a down-mix signal that can be decoded by two or
more multi-channel decoders. In other words, a compatible down-mix signal is a
down-mix signal that is initially optimized for a predetermined multi-channel decoder
and that can be converted afterwards into a signal optimized for a multi-channel
decoder, other than the predetermined multi-channel decoder, through a compatibility
processing operation.
[240] Referring to FIG. 14, assume that an input compatible down-mix signal is
optimized for the first multi-channel decoder 910. In order for the second multi-
channel decoder 930 to decode the input compatible down-mix signal, the down-mix
compatibility processing unit 920 may perform a compatibility processing operation on
the input compatible down-mix signal so that the input compatible down-mix signal
can be converted into a signal optimized for the second multi-channel decoder 930.

WO 2007/091842 PCT/KR2007/000668
The first multi-channel decoder 910 generates a first multi-channel signal by decoding
the input compatible down-mix signal. The first multi-channel decoder 910 can
generate a multi-channel signal through decoding simply using the input compatible
down-mix signal without a requirement of spatial information.
[241] The second multi-channel decoder 930 generates a second multi-channel signal
using a down-mix signal obtained by the compatibility processing operation performed
by the down-mix compatibility processing unit 920. The 3D rendering unit 940 may
generate a decoder 3D down-mix signal by performing a 3D rendering operation on the
down-mix signal obtained by the compatibility processing operation performed by the
down-mix compatibility processing unit 920.
[242] A compatible down-mix signal optimized for a predetermined multi-channel
decoder may be converted into a down-mix signal optimized for a multi-channel
decoder, other than the predetermined multi-channel decoder, using compatibility in-
formation such as an inversion matrix. For example, when there are first and second
multi-channel encoders using different encoding methods and first and second multi-
channel decoders using different encoding/decoding methods, an encoding apparatus
may apply a matrix to a down-mix signal generated by the first multi-channel encoder,
thereby generating a compatible down-mix signal which is optimized for the second
multi-channel decoder. Then, a decoding apparatus may apply an inversion matrix to
the compatible down-mix signal generated by the encoding apparatus, thereby
generating a compatible down-mix signal which is optimized for the first multi-
channel decoder.
[243] Referring to FIG. 14, the down-mix compatibility processing unit 920 may perform
a compatibility processing operation on the input compatible down-mix signal using an
inversion matrix, thereby generating a down-mix signal which is optimized for the
second multi-channel decoder 930.
[244] Information regarding the inversion matrix used by the down-mix compatibility
processing unit 920 may be stored in the decoding apparatus 900 in advance or may be
included in an input bitstream transmitted by an encoding apparatus. In addition, in-
formation indicating whether a down-mix signal included in the input bitstream is an
arbitrary down-mix signal or a compatible down-mix signal may be included in the
input bitstream.
[245] Referring to FIG. 14, the down-mix compatibility processing unit 920 includes a
first domain converter 921, a compatibility processor 922, and a second domain
converter 923.
[246] The first domain converter 921 converts the domain of the input compatible down-
mix signal into a predetermined domain, and the compatibility processor 922 performs
a compatibility processing operation using compatibility information such as an

WO 2007/091842 PCT7KR2007/000668
inversion matrix so that the input compatible down-mix signal in the predetermined
domain can be converted into a signal optimized for the second multi-channel decoder
930.
[247] The compatibility processor 922 may perform a compatibility processing operation
in a QMF/hybrid domain. For this, the first domain converter 921 may perform QMF/
hybrid analysis on the input compatible down-mix signal. Also, the first domain
converter 921 may convert the domain of the input compatible down-mix signal into a
domain, other than a QMF/hybrid domain, for example, a frequency domain such as a
DFT or FFT domain, and the compatibility processor 922 may perform the com-
patibility processing operation in a domain, other than a QMF/hybrid domain, for
example, a frequency domain or a time domain.
[248] The second domain converter 923 converts the domain of a compatible down-mix
signal obtained by the compatibility processing operation. More specifically, the
second domain converter 923 may convert the domain of the compatibility down-mix
signal obtained by the compatibility processing operation into the same domain as the
original input compatible down-mix signal by inversely performing a domain
conversion operation performed by the first domain converter 921.
[249] For example, the second domain converter 923 may convert the compatible down-
mix signal obtained by the compatibility processing operation into a time-domain
signal by performing QMF/hybrid synthesis on the compatible down-mix signal
obtained by the compatibility processing operation. Alternatively, the second domain
converter 923 may perform IDFT or IFFT on the compatible down-mix signal obtained
by the compatibility processing operation.
[250] The 3D rendering unit 940 may perform a 3D rendering operation on the
compatible down-mix signal obtained by the compatibility processing operation in a
frequency domain, a QMF/hybrid domain or a time domain. For this, the 3D rendering
unit 940 may include a domain converter (not shown). The domain converter converts
the domain of the input compatible down-mix signal into a domain in which a 3D
rendering operation is to be performed or converts the domain of a signal obtained by
the 3D rendering operation.
[251] The domain in which the compatibility processor 922 performs a compatibility
processing operation may be the same as or different from the domain in which the 3D
rendering unit 940 performs a 3D rendering operation.
[252] FIG. 15 is a block diagram of a down-mix compatibility processing/3D rendering
unit 950 according to an embodiment of the present invention. Referring to FIG. 15,
the down-mix compatibility processing/3D rendering unit 950 includes a first domain
converter 951, a second domain converter 952, a compatibility/3D rendering processor
953, and a third domain converter 954.

WO 2007/091842 PCT/KR2007/000668
[253] The down-mix compatibility processing/3D rendering unit 950 performs a com-
patibility processing operation and a 3D rendering operation in a single domain,
thereby reducing the amount of computation of a decoding apparatus.
[254] The first domain converter 951 converts an input compatible down-mix signal into
a first domain in which a compatibility processing operation and a 3D rendering
operation are to be performed. The second domain converter 952 converts spatial in-
formation and compatibility information, for example, an inversion matrix, so that the
spatial information and the compatibility information can become applicable in the
first domain.
[255] For example, the second domain converter 952 maps an inversion matrix cor-
responding to a parameter band in a QMF/hybrid domain to a frequency domain so that
the inversion matrix can become readily applicable in a frequency domain.
[256] The first domain may be a frequency domain such as a DFT or FFT domain, a
QMF/hybrid domain, or a time domain. Alternatively, the first domain may be a
domain other than those set forth herein.
[257] During the conversion of the spatial information and the compatibility information,
a time delay may occur. In order to address this problem,
[258] In order to address this problem, the second domain converter 952 may perform a
time delay compensation operation so that a time delay between the domain of the
spatial information and the compensation information and the first domain can be
compensated for.
[259] The compatibility/3D rendering processor 953 performs a compatibility processing
operation on the input compatible down-mix signal in the first domain using the
converted compatibility information and then performs a 3D rendering operation on a
compatible down-mix signal obtained by the compatibility processing operation. The
compatibility/3D rendering processor 953 may perform a compatibility processing
operation and a 3D rendering operation in a different order from that set forth herein.
[260] The compatibility/3D rendering processor 953 may perform a compatibility
processing operation and a 3D rendering operation on the input compatible down-mix
signal at the same time. For example, the compatibility/3D rendering processor 953
may generate a 3D down-mix signal by performing a 3D rendering operation on the
input compatible down-mix signal in the first domain using a new filter coefficient,
which is the combination of the compatibility information and an existing filter co-
efficient typically used in a 3D rendering operation.
[261] The third domain converter 954 converts the domain of the 3D down-mix signal
generated by the compatibility/3D rendering processor 953 into a frequency domain.
[262] FIG. 16 is a block diagram of a decoding apparatus for canceling crosstalk
according to an embodiment of the present invention. Referring to FIG. 16, the

WO 2007/091842 PC17KR2007/000668
decoding apparatus includes a bit unpacking unit 960, a down-mix decoder 970, a 3D
rendering unit 980, and a crosstalk cancellation unit 990. Detailed descriptions of the
same decoding processes as those of the embodiment of FIG. 1 will be omitted.
[263] A 3D down-mix signal output by the 3D rendering unit 980 may be reproduced by
a headphone. However, when the 3D down-mix signal is reproduced by speakers that
are distant apart from a user, inter-channel crosstalk is likely to occur.
[264] Therefore, the decoding apparatus may include the crosstalk cancellation unit 990
which performs a crosstalk cancellation operation on the 3D down-mix signal.
[265] The decoding apparatus may perform a sound field processing operation.
[266] Sound field information used in the sound field processing operation, i.e., in-
formation identifying a space in which the 3D down-mix signal is to be reproduced,
may be included in an input bitstream transmitted by an encoding apparatus or may be
selected by the decoding apparatus.
[267] The input bitstream may include reverberation time information. A filter used in
the sound field processing operation may be controlled according to the reverberation
time information.
[268] A sound field processing operation may be performed differently for an early part
and a late reverberation part. For example, the early part may be processed using a FIR
filter, and the late reverberation part may be processed using an IIR filter.
[269] More specifically, a sound field processing operation may be performed on the
early part by performing a convolution operation in a time domain using an FIR filter
or by performing a multiplication operation in a frequency domain and converting the
result of the multiplication operation to a time domain. A sound field processing
operation may be performed on the late reverberation part in a time domain.
[270] The present invention can be realized as computer-readable code written on a
computer-readable recording medium. The computer-readable recording medium may
be any type of recording device in which data is stored in a computer-readable manner.
Examples of the computer-readable recording medium include a ROM, a RAM, a CD-
ROM, a magnetic tape, a floppy disc, an optical data storage, and a carrier wave (e.g.,
data transmission through the Internet). The computer-readable recording medium can
be distributed over a plurality of computer systems connected to a network so that
computer-readable code is written thereto and executed therefrom in a decentralized
manner. Functional programs, code, and code segments needed for realizing the
present invention can be easily construed by one of ordinary skill in the art.
[271] As described above, according to the present invention, it is possible to efficiently
encode multi-channel signals with 3D effects and to adaptively restore and reproduce
audio signals with optimum sound quality according to the characteristics of a re-
production environment.

WO 2007/091842 PCT7KR2007/000668
Industrial Applicability
[272] Other implementations are within the scope of the following claims. For example,
grouping, data coding, and entropy coding according to the present invention can be
applied to various application fields and various products. Storage media storing data
to which an aspect of the present invention is applied are within the scope of the
present invention.

WO 2007/091842 PCT/KR2007/000668
Claims
[1] A decoding method of restoring a multi-channel signal, the decoding method
comprising:
extracting a three-dimensional (3D) down-mix signal and spatial information
from an input bitstream;
removing 3D effects from the 3D down-mix signal by performing a 3D rendering
operation on the 3D down-mix signal; and
generating a multi-channel signal using the spatial information and a down-mix
signal obtained by the removal.
[2] The decoding method of claim 1, wherein the removal comprises using an
inverse filter of a filter used for generating the 3D down-mix signal.
[3] The decoding method of claim 2, wherein information regarding the filter is
extracted from the input bitstream.
[4] The decoding method of claim 1, wherein the removal comprises using an
inverse function of a head related transfer function (HRTF) used for generating
the 3D down-mix signal.
[5] The decoding method of claim 4, wherein information regarding coefficients of
the HRTF or coefficients of the inverse function of the HRTF is extracted from
the input bitstream.
[6] The decoding method of claim 1, wherein the input bitstream includes at least
one of information indicating whether the input bitstream includes filter in-
formation identifying a filter used to perform the 3D rendering operation and in-
formation indicating whether the filter information specifies an inverse filter of a
filter used for generating the 3D down-mix signal
[7] The decoding method of claim 1, wherein the removal comprises performing the
3D rendering operation in one of a discrete Fourier transform (DFT) domain, a
fast Fourier transform (FFT) domain, a quadrature mirror filter (QMF)/hybrid
domain, and a time domain.
[8] The decoding method of claim 1, further comprising decoding the 3D down-mix
signal.
[9] A decoding method of restoring a multi-channel signal, the decoding method
comprising:
extracting a 3D down-mix signal and spatial information from an input bitstream;
generating a multi-channel signal using the 3D down-mix signal and the spatial
information; and
removing 3D effects from the multi-channel signal by performing a 3D rendering
operation on the multi-channel signal.

WO 2007/091842 PCT7KR2007/000668
[10] An encoding method of encoding a multi-channel signal with a plurality of
channels, the encoding method comprising:
encoding the multi-channel signal into a down-mix signal with fewer channels;
generating spatial information regarding the plurality of channels;
generating a 3D down-mix signal by performing a 3D rendering operation on the
down-mix signal; and
generating a bitstream including the 3D down-mix signal and the spatial in-
formation.
[11] The encoding method of claim 10, wherein the generation of the 3D down-mix
signal comprises performing the 3D rendering operation using a HRTF.
[12] The encoding method of claim 11, wherein the bitstream includes at least one of
information regarding coefficients of the HRTF and information regarding co-
efficients of an inverse function of the HRTF.
[13] The encoding method of claim 10, wherein the generation of the 3D down-mix
signal comprises performing the 3D rendering operation in one of a DFT domain,
an FFT domain, a QMF/hybrid domain, and a time domain.
[14] An encoding method of encoding a multi-channel signal with a plurality of
channels, the encoding method comprising:
performing a 3D rendering operation on the multi-channel signal;
encoding a multi-channel signal obtained by the 3D rendering operation into a
3D down-mix signal with fewer channels;
generating spatial information regarding the plurality of channels; and
generating a bitstream including the 3D down-mix signal and the spatial in-
formation.
[15] A decoding apparatus for restoring a multi-channel signal, the decoding
apparatus comprising:
a bit unpacking unit which extracts an encoded 3D down-mix signal and spatial
information from an input bitstream;
a down-mix decoder which decodes the encoded 3D down-mix signal;
a 3D rendering unit which removes 3D effects from the decoded 3D down-mix
signal obtained by the decoding performed by the down-mix decoder by
performing a 3D rendering operation on the decoded 3D down-mix signal; and
a multi-channel decoder which generates a multi-channel signal using the spatial
information and a down-mix signal obtained by the removal performed by the
3D rendering unit.
[16] A decoding apparatus for restoring a multi-channel signal, the decoding
apparatus comprising:
a bit unpacking unit which extracts an encoded 3D down-mix signal and spatial

WO 2007/091842 PCT7KR2007/000668
information from an input bitstream;
a down-mix decoder which decodes the encoded 3D down-mix signal;
a multi-channel decoder which generates a multi-channel signal using the spatial
information and a 3D down-mix signal obtained by the decoding performed by
the down-mix decoder; and
a 3D rendering unit which removes 3D effects from the multi-channel signal by
performing a 3D rendering operation on the multi-channel signal.
[17] An encoding apparatus for encoding a multi-channel signal with a plurality of
channels, the encoding apparatus comprising:
a multi-channel encoder which encodes the multi-channel signal into a down-mix
signal with fewer channels and generates spatial information regarding the
plurality of channels;
a 3D rendering unit which generates a 3D down-mix signal by performing a 3D
rendering operation on the down-mix signal;
a down-mix encoder which encodes the 3D down-mix signal; and
a bit packing unit which generates a bitstream including the encoded 3D down-
mix signal and the spatial information.
[18] An encoding apparatus for encoding a multi-channel signal with a plurality of
channels, the encoding apparatus comprising:
a 3D rendering unit which performs a 3D rendering operation on the multi-
channel signal;
a multi-channel encoder which encodes a multi-channel signal obtained by the
3D rendering operation into a 3D down-mix signal with fewer channels and
generates spatial information regarding the plurality of channels;
a down-mix encoder which encodes the 3D down-mix signal; and
a bit packing unit which generates a bitstream including the encoded 3D down-
mix signal and the spatial information.
[19] A computer-readable recording medium having a computer program for
executing the decoding method of any one of claims 1 through 9 or the encoding
method of any one of claims 10 through 14.
[20] A bitstream comprising:
a data field which includes information regarding a 3D down-mix signal;
a filter information field which includes filter information identifying a filter
used for generating the 3D down-mix signal;
a first header field which includes information indicating whether the filter in-
formation field includes the filter information;
a second header field which includes information indicating whether the filter in-
formation field includes coefficients of the filter or coefficients of an inverse

WO 2007/091842 PCT/KR2007/000668
filter of the filter; and
a spatial information field which includes spatial information regarding a
plurality of channels.

An encoding method and apparatus and a decoding method and apparatus are provided. The decoding method includes extracting a three-dimensional (3D) down-mix signal and spatial information from an input bitstream, removing 3D effects from the 3D down-mix signal by performing a 3D rendering operation on the 3D down-mix signal, and generating a multi-channel signal using the spatial information and a down-mix signal obtained by the removal. Accordingly, it is possible to efficiently encode multi-channel signals with 3D effects and to adaptively restore and reproduce audio signals with optimum sound quality according to the characteristics of a reproduction environment.

Documents:

http://ipindiaonline.gov.in/patentsearch/GrantedSearch/viewdoc.aspx?id=ELE0PE/SgPJ/fK5cmqP/BQ==&loc=wDBSZCsAt7zoiVrqcFJsRw==


Patent Number 271384
Indian Patent Application Number 2897/KOLNP/2008
PG Journal Number 08/2016
Publication Date 19-Feb-2016
Grant Date 18-Feb-2016
Date of Filing 16-Jul-2008
Name of Patentee LG ELECTRONICS INC.
Applicant Address 20, YOIDO-DONG, YOUNGDUNGPO-GU, SEOUL
Inventors:
# Inventor's Name Inventor's Address
1 JUNG, YANG WON # 2-803, YEOKSAM HANSIN APT., DOGOK-DONG, GANGNAM-KU, SEOUL 135-720
2 OH, HYUN O #306-403, HANSIN APT., GANGSUN VILLAGE 3DANJI, JUYEOP1-DONG, IISAN SEO-KU, GOYANG-SI, KYUNGGI-DO, 411-744
3 KIM, DONG SOO #502, WOOLIM VILLA, 602-265, NAMHYUN-DONG, KWANAK-KU, SEOUL, 151-080
4 LIM, JAE HYUN # 609, PARKVILLE OFFICETEL, 1062-20, NAMHYUN-DONG, KWANAK-KU, SEOUL 151-080
5 PANG, HEE SUK #101, 4/7, 14-10, YANGJAE-DONG, SEOCHO-KU, SEOUL 137-130
PCT International Classification Number G10L 19/00
PCT International Application Number PCT/KR2007/000668
PCT International Filing date 2007-02-07
PCT Conventions:
# PCT Application Number Date of Convention Priority Country
1 60/773337 2006-02-15 U.S.A.
2 60/792329 2006-04-17 U.S.A.
3 60/782519 2006-03-16 U.S.A.
4 60/781750 2006-03-14 U.S.A.
5 60/765747 2006-02-07 U.S.A.