The present disclosure relates to a device and a method for playing back an audio object using one or more speaker arrays. The present disclosure relates particularly to a device and a method for playing back an audio object including playback position information indicating a position at which a sound image is to be localized in a three-dimensional space.
In recent years, many digital television broadcast receivers and DVD players for playing back 5.1ch audio content items have been developed and prepared for the market. Here, “5.1ch” is a channel setting for arranging front left and right channels, a front center channel, and left and right surround channels. Some of recent Blu-ray (registered trademark) players have a 7.1ch configuration in which left and right back surround channels are added.
On the other hand, with further increases in the sizes of image screens and in the definitions of images, virtual surround of audio objects has been vigorously studied. For example, virtual surround in the case where 22.2ch speakers are arranged has been studied.
In addition, effort for differentiating movie theaters using three-dimensional acoustic effects have been vigorously made (Non-patent Literature 2). In this case, speakers are arranged also on a ceiling in a three-dimensional (3D) configuration. Here, content items are coded as audio objects. An audio object is an audio signal with playback position information indicating, in a three-dimensional space, the position at which a sound image is to be localized. For example, an audio object is a coded signal of a pair of (i) playback position information indicating the position at which a sound source (sound image) is localized in the form of coordinates (x, y, z) along three axes and (ii) an audio signal of the sound source.
For example, when creating an audio object of any of a bullet, an airplane, and a note of a flying bird, etc., the position indicated by playback position information is caused to transit with time from one minute to the next. In this case, the playback position information may be vector information indicating a transition direction. In the case of an explosion sound etc. generated at a certain position, playback position information is naturally constant.
In this way, playback of audio signals with playback position information has been researched and developed on the premise that speakers are arranged three-dimensionally. However, it is impossible to arrange speakers three-dimensionally in many cases for actual home use or personal use.
As a technique for enabling audio playback with higher-possible realistic sensations under an environment where speakers cannot be arranged freely, a method using a head related transfer function (HRTF), wavefront synthesis, and beam forming, etc. have been researched and developed.
The HRTF is a transfer function for simulating propagation property of a sound around the head of a listener. A perception of a sound arrival direction is said to be affected by the HRTF. As illustrated in
The basic operation principle of the wavefront synthesis is as illustrated in (a) of
The basic operation principle of the beam forming is as illustrated in (b) of
International Publication No. 2006/030692
First published in SMPTE Technical Conference Publication in October, 2007
Dolby Atmos Cinema Technical Guidelines
Audio Eng Soc, Vol 49, No 4, 2001 April Introduction to Head-Related Transfer Functions (HRTFs): Representations of HRTFs in Time, Frequency, and Space
Audio Signal Processing for Next-Generation Multimedia Communication Systems, pp. 323-342, Y. A. Huang, J. Benesty, Kluwer, January 2004
AES 127th Convention, New York N.Y., USA, 2009, Oct. 9-12 Physical and Perceptual Properties of Focused Sources in Wave Field Synthesis
There is a problem that it is difficult to produce, in actual home use or personal use, a configuration in which speakers are arranged on a ceiling as in the 22.2ch configuration described above.
Methods for providing highly realistic sound even in the case where speakers cannot be freely arranged include the method using an HRTF, the wavefront synthesis, and beam forming. The method using an HRTF is excellent as a method for controlling a sound arrival direction, but does not reproduce any sensation of distance between a listener and a sound source because the method using an HRTF is merely for performing control for creating the acoustic signal that perceptually sounds from the direction and thus does not reproduce actual physical wavefronts. On the other hand, the wavefront synthesis and the beam forming can reproduce actual physical wavefronts, and thus can reproduce a sensation of distance between the listener and the sound source, but cannot generate the sound source behind the listener. This is because the sound waves output from the speaker array reach the ears of the listener before the sound waves form a sound image.
In addition, since each of the conventional techniques is a technique for controlling a sound on the two-dimensional plane on which the speakers are arranged, it is impossible to perform signal processing reflecting playback position information when the playback position information included in the audio object is represented as three-dimensional space information.
The present disclosure has been made in view of the conventional problems, and has an object to provide an audio playback device and an audio playback method for playing back an audio object including three-dimensional playback position information with highly realistic sensations even in a space where speakers cannot be arranged freely.
In order to solve the above-described problems, an audio playback device according to an embodiment is an audio playback device which plays back an audio object including an audio signal and playback position information indicating a position in a three-dimensional space at which a sound image of the audio signal is localized, the audio playback device including: at least one speaker array which converts an acoustic signal to acoustic vibration; a converting unit configured to convert the playback position information to corrected playback position information which is information indicating a position of the sound image on a two-dimensional coordinate system based on a position of the at least one speaker array; and a signal processing unit configured to localize the sound image of the audio signal included in the audio object according to the corrected playback position information.
With this configuration, since the three-dimensional playback position information included in the audio object is converted into the corrected playback position information on the two-dimensional coordinate system based on the position of the at least one speaker array, and the sound image is localized according to the corrected playback position information, it is possible to play back the audio object with highly realistic sensations even when there is a restriction on the arrangement of the at least one speakers.
Here, when (i) a direction in which speaker elements are arranged in each of the at least one speaker array is an X axis, (ii) a direction which is orthogonal to the X axis and parallel to a setting surface on which the at least one speaker array is arranged is a Y axis, and (iii) a direction which is orthogonal to the X axis and perpendicular to the setting surface is a Z axis, the corrected playback position information may indicate the position at coordinates (x, y) on the two-dimensional coordinate system expressed by the X axis and the Y axis, and when the position identified by the playback position information is expressed by coordinates (x, y, z), the corrected playback position information may indicate values corresponding to x and y.
In this case, since the corrected playback position information indicates values according to the x-coordinate value and the y-coordinate value when the position identified by the playback position information is expressed by (x, y, z), it is possible to play back the audio object including the three-dimensional playback position information with highly realistic sensations even in a space where the speakers cannot be arranged three-dimensionally.
In addition, when, on the two-dimensional coordinate system, (i) a y coordinate located behind the speaker array is a negative coordinate and a y coordinate located in front of the speaker array is a positive coordinate, and (ii) an x coordinate located to a left of a center of the speaker array is a negative coordinate and an x coordinate located to a right of the center of the speaker array is a positive coordinate, a value of the corrected playback position information may be a value obtained by multiplying at least one of the x-coordinate value and the y-coordinate value by a predetermined value.
In this case, since the values of the corrected playback position information are obtained by multiplying the at least one of the coordinates (x, y) by the predetermined value, the recognizable size of the area can be virtually changed.
In addition, an x-coordinate value of the corrected playback position information may be limited to a width of the at least one speaker array.
In this case, the x-coordinate value of the corrected playback position information is a value limited to the width of the at least one speaker array, it is possible to perform signal processing suitable for the performance of the at least one speaker array.
In addition, the signal processing unit may be a beam forming unit configured to form a sound image at the position on the two-dimensional coordinate system.
In this case, since strong acoustic vibration is generated by the beam forming unit at a target position, it is possible to generate a sound field in which a sound source is virtually present at the target position.
In addition, when, on the two-dimensional coordinate system, a y coordinate located behind the speaker array is a negative coordinate and a y coordinate located in front of the speaker array is a positive coordinate, and the signal processing unit may be configured to perform wavefront synthesis by signal processing using a Huygens' principle when a y-coordinate value of the corrected playback position information is a negative value.
In this case where the y-coordinate value of the corrected playback position information is the negative value, wavefront synthesis is performed by signal processing using the Huygens' principle. Thus, it is possible to generate a sound field in which a sound source is virtually present at the target position even when the target position of the sound image to be localized is behind the speakers.
In addition, the corrected playback position information may indicate the position on the two-dimensional coordinate system, the position being indicated by (i) a direction angle to the position indicated by the playback position information when seen from a position of a listener listening to an acoustic sound output from the at least one speaker array and (ii) a distance from the position of the listener to the position indicated by the playback position information.
In this way, since the corrected playback position information indicates the position on the two-dimensional coordinate system in the form of the direction angle to the position indicated by the playback position information when seen from the position of the listener and the distance from the position of the listener to the position indicated by the playback position information. Thus, it is possible to control the virtually sensible direction in which the sound source is present with respect to the position of the listener and the virtually sensible distance from the position of the listener to the sound source.
In addition, the signal processing unit may be configured to localize the sound image using a head related transfer function (HRTF), and the HRTF may be set so that a sound may be audible from a direction of the position indicated by the corrected playback position information.
In this case, since the sound image is localized using the HRTF so that the sound is audible from the direction of the position indicated by the corrected playback position information, it is possible to perform playback reflecting the direction to the sound source when the sound is listened to by the listener.
In addition, the signal processing unit may be configured to adjust a sound volume according to the distance from the position of the listener to the position indicated by the corrected playback position information.
In this case, since the sound volume is adjusted according to the distance between the position of the listener and the position indicated by the corrected playback position information, it is possible to perform playback reflecting the distance to the sound source when the sound is listened to by the listener.
In addition, the signal processing unit may be configured to change a signal processing method according to the position indicated by the corrected playback position information.
In this case, since the signal processing method is changed according to the position indicated by the corrected playback position information, it is possible to select an optimum signal processing method according to the target playback position.
In addition, when (i) a direction in which speaker elements are arranged in each of the at least one speaker array is an X axis, (ii) a direction which is orthogonal to the X axis and parallel to a setting surface on which the at least one speaker array is arranged is a Y axis, and (iii) a direction which is orthogonal to the X axis and perpendicular to the setting surface is a Z axis, when, on the two-dimensional coordinate system, a y coordinate located behind the speaker array is a negative coordinate and a y coordinate located in front of the speaker array is a positive coordinate, the signal processing unit may be configured to: when a y-coordinate value of the corrected playback position information is a negative value, perform wavefront synthesis by signal processing using a Huygens' principle; when a y-coordinate value of the corrected playback position information is a positive value indicating a position in front of a listener, generate a sound image by signal processing using beam forming; and when a y-coordinate value of the corrected playback position information is a positive value indicating a position behind the listener, localize a sound image by signal processing using a head related transfer function (HRTF).
In this case, the signal processing unit (i) performs the wavefront synthesis by signal processing using the Huygens' principle when the y-coordinate value of the corrected playback position information is the negative value, (ii) generates the sound image by signal processing using the beam forming when the y-coordinate value of the corrected playback position information is the positive value indicating the position in front of the listener, and (iii) localizes the sound image by signal processing by using the HRTF when the y-coordinate value of the corrected playback position information is the positive value indicating the position behind the listener. Thus, it is possible to create a sound field where the acoustic vibration is generated and virtually presented at the target position in front of the position of the listener and to perform playback in the sound field where a sound virtually and perceptually approaches from the direction behind the position of the listener.
In addition, the audio playback device may include at least two speaker arrays, wherein each of the at least two speaker arrays forms a corresponding one of at least two two-dimensional coordinate systems, and when the position identified by the playback position information is expressed by coordinates (x, y, z) where (i) a direction in which speaker elements are arranged in one of the at least two speaker arrays is an X axis, (ii) a direction which is orthogonal to the X axis and parallel to a setting surface on which the one of the at least two speaker arrays is arranged is a Y axis, and (iii) a direction which is orthogonal to the X axis and perpendicular to the setting surface is a Z axis, the signal processing unit may be configured to control the at least two speaker arrays according to a z-coordinate value. When the two two-dimensional coordinate systems are parallel to each other, the signal processing unit may be configured to: increase a sound volume of the one of the at least two speaker arrays which is on an upper two-dimensional coordinate system with respect to the setting surface when the z-coordinate value is larger than a predetermined value; and increase a sound volume of the one of the at least two speaker arrays which is on a lower two-dimensional coordinate system with respect to the setting surface when the z-coordinate value is smaller than the predetermined value. When the two two-dimensional coordinate systems are orthogonal to each other, the signal processing unit may be configured to: increase a sound volume of one or more speaker elements in the one of the at least two speaker arrays when the z-coordinate value is larger than a predetermined value, the one or more speaker elements being arranged at positions above a predetermined position on a two-dimensional coordinate system perpendicular to the setting surface among the at least two two-dimensional coordinate systems; and increase a sound volume of one or more speaker elements in the one of the at least two speaker arrays when z-coordinate value is smaller than the predetermined value, the one or more speaker elements being arranged at positions below the predetermined position on the two-dimensional coordinate system perpendicular to the setting surface among the at least two two-dimensional coordinate systems.
In this way, the audio playback device includes the at least two speaker arrays which are controlled according to the value of z in coordinates (x, y, z) indicating the position identified by the playback position information. Thus, it is possible to control the height information of the playback position information, and to play back the audio object including the three-dimensional playback position information with highly realistic sensations.
In addition, an audio playback device according to an embodiment is an audio playback device which plays back an audio object including an audio signal and playback position information indicating a position in a three-dimensional space at which a sound image of the audio signal is localized, wherein the audio object includes an audio frame including the audio signal which is obtained at a predetermined time interval and the playback position information, and when the playback position information of the audio frame included in the audio object is lost, the audio playback device plays back the audio frame by using playback position information included in an audio frame that has been played back previously as playback position information of the audio frame whose playback position information is lost.
In this way, when the playback position information of the current audio frame is lost, the playback position information included in the audio frame that has been previously played back is used. Thus, even when the playback position information of the current audio frame is lost, it is possible to create a natural sound field, or to reduce the amount of information required to record or transmit the audio object when the audio object is not moving.
It is to be noted that other possible embodiments for solving the problems include not only the audio playback device described above but also an audio playback method, a program for executing the audio playback method, and a computer-readable recording medium such as a DVD on which the program is recorded.
The audio playback device and the audio playback method make it possible to play back an audio object including three-dimensional playback position information with highly realistic sensations even in a space in which speakers cannot be freely arranged.
These and other objects, advantages and features of the disclosure will become apparent from the following description thereof taken in conjunction with the accompanying drawings that illustrate a specific embodiment of the present disclosure.
Hereinafter, an embodiment of an audio playback device and an audio playback method is described with reference to the drawings.
It is to be noted that the embodiment described below indicates a preferred specific example. The numerical values, shapes, constituent elements, the arrangement and connection of the constituent elements, the processing order of operations etc. indicated in the following embodiment are mere examples, and therefore do not limit the scope of the present disclosure. Therefore, among the constituent elements in the following embodiment, constituent elements not recited in any one of the independent claims that define the most generic concept of the present disclosure are described as arbitrary constituent elements.
In
The setting unit 101 is a processing unit which sets a virtual two-dimensional coordinate system according to a position at which the speaker array 106 is arranged (the two-dimensional coordinate system is determined based on the position of the speaker array 106).
The converting unit 102 is a processing unit which converts the playback position information obtained by the audio object dividing unit 100 into corrected playback position information which is position information (two-dimensional information) on the two-dimensional coordinate system set by the setting unit 101.
The selecting unit 103 is a processing unit which selects a signal processing method that should be employed by the signal processing unit 105, based on the corrected playback position information generated by the converting unit 102; the two-dimensional coordinate system set by the setting unit 101; and the position of a listener listening to an acoustic sound output from the speaker array 106 (the position predetermined by the audio playback device 110).
The decoding unit 104 is a processing unit which decodes the coded audio signal obtained by the audio object dividing unit 100 to generate an audio signal (acoustic signal).
The signal processing unit 105 is a processing which localizes a sound image of the audio signal obtained through the decoding by the decoding unit 104 according to the corrected playback position information obtained through the conversion by the converting unit 102. Here, the signal processing unit 105 performs the processing according to the signal processing method selected by the selecting unit 103.
The speaker array 106 is at least one speaker array (a group of speaker elements arranged in a column) which converts an output signal (the acoustic signal) from the signal processing unit to acoustic vibration.
The audio object dividing unit 100, the setting unit 101, the converting unit 102, the selecting unit 103, the decoding unit 104, the signal processing unit 105 are typically implemented as hardware using electronic circuits such as semiconductor integrated circuits, and alternatively may be implemented as software using one or more programs each executable by a computer including a CPU, a ROM, a RAM, or the like.
Hereinafter, descriptions are given of operations performed by the thus-configured audio playback device 110 according to this embodiment.
First, the audio object dividing unit 100 divides the audio object including the playback position information and the coded audio signal into the playback position information and the coded audio signal. For example, the audio object has a configuration as illustrated in
The audio object dividing unit 100 extracts the playback position information and the coded audio signal from the audio object as illustrated in
The setting unit 101 sets a virtual two-dimensional coordinate system according to the position at which the speaker array 106 is arranged. A schematic view of the speaker array 106 is illustrated in
Next, the converting unit 102 converts the three-dimensional playback position information into corrected playback position information which is two-dimensional information. In this embodiment, a two-dimensional coordinate system having the X axis and the Y axis as illustrated in each of
The conversion from the three-dimensional playback position information to the two-dimensional corrected position information is performed, for example, according to one of methods illustrated in
One of methods illustrated in
In
The signal processing unit 105 may perform processing for localizing a sound image according to the method using an HRTF set so that sound is audible from the direction of the position indicated by the corrected playback position information. In this way, it is possible to control the virtually sensible direction in which the sound source is present with respect to the position of the listener and the virtually sensible distance from the position of the listener to the sound source. Furthermore, the signal processing unit 105 may adjust a sound volume according to the distance (r1′, r2′, etc.) from the position of the listener and the position indicated by the corrected playback position information. In this way, it is possible to perform playback reflecting the virtually sensible distance from the listener to the sound source.
Next, the selecting unit 103 selects the signal processing method that should be employed by the signal processing unit 105 based on (i) the corrected playback position information generated by the converting unit 102, (ii) the two-dimensional coordinate system set by the setting unit 101, and (iii) the position of the listener (or the listener's listening position predetermined by the audio playback device 110).
On the other hand, the coded audio signal obtained by the audio object dividing unit 100 is decoded into an audio PCM signal by the decoding unit 104. The decoding unit 104 may be any decoder conforming to a codec method used to code the coded audio signal.
The audio PCM signal decoded in this way is processed by the signal processing unit 105 according to the signal processing method selected by the selecting unit 103. More specifically, the signal processing unit 105 (i) performs the wavefront synthesis by signal processing using the Huygens' principle when the y-coordinate value of the corrected playback position information is a negative value, (ii) generates a sound image by signal processing using the beam forming when the y-coordinate value of the corrected playback position information is a positive value indicating a position in front of the listener, and (iii) localizes a sound image by signal processing according to the method using an HRTF when the y-coordinate value of the corrected playback position information is a positive value indicating a position behind the listener.
In this embodiment, the signal processing method is any one of the beam forming, the wavefront synthesis, and the method using an HRTF. Any of the signal processing methods can be specifically performed using a conventional signal processing method.
Lastly, the speaker array 106 converts the output signal (acoustic signal) from the signal processing unit 105 into acoustic vibration.
First, the audio object dividing unit 100 divides an audio object into three-dimensional playback position information and a coded audio signal (S10).
Next, the converting unit 102 converts the three-dimensional playback position information obtained by the audio object dividing unit 100 into corrected playback position information which is position information (two-dimensional information) on the two-dimensional coordinate system based on the position of the speaker array 106 (S11).
Next, the selecting unit 103 selects a signal processing method that should be employed by the signal processing unit 105, based on the corrected playback position information generated by the converting unit 102; the two-dimensional coordinate system set by the setting unit 101; and the position of a listener listening to an acoustic sound output from the speaker array 106 (the position may be a listener's position predetermined by the audio playback device 110) (S12).
Lastly, the signal processing unit 105 localizes the sound image of the audio signal obtained by the audio object dividing unit 100 and then decoded by the decoding unit 104, according to the corrected playback position information obtained through the conversion by the converting unit 102 (S13). At this time, the signal processing unit 105 performs the processing using the signal processing method selected by the selecting unit 103.
In this way, the three-dimensional playback position information included in the audio object is converted into the corrected playback position information on the two-dimensional coordinate system based on the position of the speaker array, and the sound image is localized according to the corrected playback position information. Thus, even when there is a restriction on the arrangement of the speaker array, the audio object can be played back with highly realistic sensations.
Alternatively, in addition to the steps S10 to S13 illustrated in
The audio object dividing unit 100 determines whether playback position information of a current audio frame is lost (S20).
When it is determined that the playback position information is lost (Yes in S20), playback position information included in an audio frame that has been previously played back is used by the audio object dividing unit 100 as a replacement for the playback position information of the current audio frame, and signal processing is performed by the signal processing unit 105 according to the playback position information (after conversion to two-dimensional corrected playback position information) (S21).
When it is determined that the playback position information is not lost (No in S20), playback position information included in a current audio frame is divided by the audio object dividing unit 100, and signal processing is performed by the signal processing unit 105 according to the playback position information (after conversion to two-dimensional corrected playback position information) (S22).
In this way, since the playback position information included in the audio frame that has been previously played back is used even when the playback position information of the current audio frame is lost, it is possible to naturally play back a sound in a sound field, or to reduce the amount of information required to record or transmit the audio object when the audio object does not move.
It is to be noted that the procedures according to the flowcharts of
In this embodiment, one of the three signal processing methods is selected according to the corrected playback position information. In
Alternatively, although one of the three signal processing methods is selected according to the corrected playback position information in this embodiment, the method using an HRTF may be selected irrespective of the position of the corrected playback position information. The method using an HRTF can be selected in any of the cases because it enables control in any of the cases by simulating binaural phase difference information, binaural level difference information, and an acoustic transfer function around the head of the listener. On the other hand, the wavefront synthesis using the Huygens' principle does not enable localization of a sound image in front of the speaker array, and the beam forming does not enable localization of a sound image behind the speaker array and behind the listener.
Although control in a Z-axis direction is not described in this embodiment, it is also possible to add the control to the method using an HRTF utilizing the result of study (Patent Literature 1) that a clue for localization in a perpendicular direction is included in an amplification spectrum of an acoustic transfer function around the head of the listener.
Alternatively, control in a Z-axis direction may be performed by creating a plurality of coordinate planes using a plurality of speaker arrays.
In each of the examples in
In another case where two two-dimensional coordinate systems are orthogonal to each other, the signal processing unit 105 increases the sound volume of one or more speaker elements in the one of the at least two speaker arrays when the value of z is larger than (or no smaller than) a predetermined value, the one or more speaker elements being arranged at positions above a predetermined position on a two-dimensional coordinate system perpendicular to the X-Y plane (setting surface) among the at least two two-dimensional coordinate systems, and increases the sound volume of one or more speaker elements in the one of the at least two speaker arrays when the value of z is smaller than (or no larger than) the predetermined value, the one or more speaker elements being arranged at positions below the predetermined position on the two-dimensional coordinate system perpendicular to the X-Y plane (setting surface) among the at least two two-dimensional coordinate systems.
In this way, when the audio playback device 110 includes at least two speaker arrays, since the at least two speaker arrays are controlled according to the value of z in coordinates (x, y, z) indicating the position identified by the playback position information, height information of the playback position information can be controlled, and the audio object including the three-dimensional playback position information can be played back with highly realistic sensations.
As described above, the audio playback device 110 in this embodiment includes: the at least one speaker array 106 which converts an acoustic signal into acoustic vibration; the converting unit 102 which converts the three-dimensional playback position information into position information (corrected playback position information) based on the position of the speaker array 106 on the two-dimensional coordinate system; and the signal processing unit 105 which localizes the sound image of the audio object according to the corrected playback position information. Thus, the audio playback device 110 is capable of playing back the audio object with the three-dimensional playback position information with optimum realistic sensations even in an environment where speakers cannot be freely arranged, specifically, no speaker can be set on a ceiling.
Although the audio playback devices according to aspects of the present invention has been described above based on the embodiment and variations thereof, audio playback devices disclosed herein are not limited to the embodiment and variations thereof. The present disclosure covers various modifications that a person skilled in the art may conceive and add to the exemplary embodiment or any of the variations or embodiments obtainable by arbitrarily combining different embodiments based on the present disclosure.
Although the setting unit 101 is included in this embodiment, the setting unit 101 is unnecessary when the setting position of the speaker array is determined in advance.
Although listener position information is input to the selecting unit 103 in this embodiment, the listener position information does not need to be input when the position of the listener is determined in advance, or the position determined in advance by the device is fixed.
The selecting unit 103 is also unnecessary when a signal processing method is fixed (for example, it is determined that processing is always performed according to the method using an HRTF).
Although the decoding unit 104 is included in this embodiment, the decoding unit 104 is unnecessary when the coded audio signal is a simple PCM signal, in other words, the audio signal included in the audio object is not coded.
Although the audio object dividing unit 100 is included in this embodiment, the audio object dividing unit 100 is unnecessary when an audio object having a structure in which an audio signal and playback position information are divided is input to the audio playback device 110.
In addition, speaker elements do not always need to be arranged linearly in the speaker array, and may be arranged in an arch (arc) shape, for example. The intervals between speaker elements do not always need to be equal. The present disclosure does not limit the shape of each of speaker arrays.
The audio playback device according to the present disclosure has one or more speaker arrays, and is particularly capable of playing back an audio object including three-dimensional position information with highly realistic sensations even in a space in which speakers cannot be arranged three-dimensionally. Thus, the audio playback device is widely applicable to devices for playing back audio signals.
Number | Date | Country | Kind |
---|---|---|---|
2013-122254 | Jun 2013 | JP | national |
This is a continuation application of PCT International Application No. PCT/JP2014/000868 filed on Feb. 19, 2014, designating the United States of America, which is based on and claims priority of Japanese Patent Application No. 2013-122254 filed on Jun. 10, 2013. The entire disclosures of the above-identified applications, including the specifications, drawings and claims are incorporated herein by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
20050105442 | Melchior et al. | May 2005 | A1 |
20060098830 | Roeder | May 2006 | A1 |
20060120533 | Chen | Jun 2006 | A1 |
20080008326 | Reichelt et al. | Jan 2008 | A1 |
20080013746 | Reichelt et al. | Jan 2008 | A1 |
20080226084 | Konagai | Sep 2008 | A1 |
20090034772 | Iida et al. | Feb 2009 | A1 |
20110144783 | Reichelt et al. | Jun 2011 | A1 |
20120070021 | Yoo | Mar 2012 | A1 |
20140064517 | Ko | Mar 2014 | A1 |
Number | Date | Country |
---|---|---|
1826838 | Aug 2006 | CN |
2001-197598 | Jul 2001 | JP |
2006-128818 | May 2006 | JP |
2011-035784 | Feb 2011 | JP |
2011-066868 | Mar 2011 | JP |
2006030692 | Mar 2006 | WO |
2013006338 | Jan 2013 | WO |
Entry |
---|
International Search Report (ISR) issued in International Application No. PCT/JP2014/000868 dated Mar. 18, 2014 with English translation. |
K. Hamasaki et aL, “Multichannel Sound System for Ultra High-Definition TV”, first published in SMPTE Technical Conference Publication, Oct. 2007. |
Dolby Atmos Cinema Technical Guidelines, pp. 1-20. |
C. Cheng et aL, “Introduction to Head-Related Transfer Functions (HRTFs): Representations of HRTFs in Time, Frequency, and Space”, J. Audio Eng. Soc., vol. 49, No. 4, Apr. 2001, pp. 231-249. |
Y. A. Huang et al., Audio Signal Processing for Next-Generation Multimedia Communication Systems, Jan. 2004, pp. 323-345. |
S. Spors et al., “Physical and Perceptual Properties of Focused Sources in Wave Field Synthesis”, Audio Engineering Society Convention Paper, presented at the 127th Convention Oct. 9-12, 2009, pp. 1-19. |
Chinese Office Action issued in Application No. 201480032404.7 dated Aug. 1, 2016, with partial English translation. |
Number | Date | Country | |
---|---|---|---|
20160088393 A1 | Mar 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/JP2014/000868 | Feb 2014 | US |
Child | 14961739 | US |