Extracting a common signal from multiple audio signals

Information

  • Patent Grant
  • 9820073
  • Patent Number
    9,820,073
  • Date Filed
    Wednesday, May 10, 2017
    8 years ago
  • Date Issued
    Tuesday, November 14, 2017
    7 years ago
  • Inventors
  • Original Assignees
  • Examiners
    • Tsang; Fan
    • Zhao; Eugene
    Agents
    • Renner, Otto, Boisselle & Sklar, LLP
Abstract
Extracting a common signal from multiple audio signals may include summing a first signal and a second signal to obtain a first+second signal; subtracting the second signal from the first signal to obtain a first−second signal; transforming the first+second signal and the first−second signal to frequency domain representations; calculating absolute value of the frequency domain representations of the first+second signal and the first−second signal; subtracting the absolute value of the frequency domain representation of the first−second signal from the absolute value of the frequency domain representation of the first+second signal to obtain a difference signal; multiplying the difference signal by the frequency domain representation of the first+second signal to obtain a product signal; dividing the product signal by the absolute value of the frequency domain representation of the first+second signal to obtain a frequency domain representation of the common signal; and transforming the frequency domain representation to the common signal.
Description
BACKGROUND

Surround audio systems used in movie theatres and home cinema systems use multiple speakers to simulate a sound field surrounding the listener. Surround audio greatly enhances the listener's experience. Currently, one of the most popular surround audio configurations is the well-known surround 5.1 configuration which uses five full bandwidth channels and a low frequency effect (LFE) channel. The most popular format for storing high quality music, however, remains two-channel stereo, not surround. Therefore, for the most part, high quality music cannot be enjoyed in surround.


Upmixing processes may be applied to audio signals to derive additional audio channels, for example, to go from two-channel stereo to surround 5.1. Existing upmixing methods create an allusion of surround sound through the use of matrixing, phasing, time delay, and reverberation effects added to the stereo mix as a way to generate surround. While these conventional systems and methods create an effect, they also produce unnatural-sounding artifacts that may cause music to sound fake, hollow, or annoying to the listener. The listener may experience a loss of overall clarity or quality as if “something just does not sound correct” with the sound. There may be little or no distinction to the segments in the conventional upmixed surround presentation. These issues diminish the listener's experience.


Another area in which sonic artifacts negatively affect the listener's experience is peak limiting. Peak limiters have become ubiquitous in music production. With the advent of digital processing, peak limiters can react instantaneously to or even in advance of audio peaks, ensuring that the signal never gets louder than a predetermined ceiling. Conventional peak limiters, however, come at the cost of clarity and punch, and added spectral artifacts. Thus, conventional peak limiters may do a good job at limiting peaks but they add harmonic spectra that disturbs the sound. These issues also diminish the listener's experience.


SUMMARY

The present disclosure describes novel techniques for extracting a common signal from multiple audio signals and for using the extracted common signal to improve the surround effect or the peak limited sound the listener hears, thereby improving the user experience.


The techniques disclosed herein include an audio signal processing technique that extracts a common signal from those embedded in a stereophonic signal. This technique is used to convert the stereophonic audio (i.e., a two channel signal) into multiple channels of surround sound. A key resulting feature of the techniques disclosed is significantly improved distinction in the sound field. The techniques isolate and place audio segments into proper sound field locations. Importantly, the techniques accomplish this without adding any fake or false signal cues, such as time delay, reverb, or ancillary unnatural effects to generate the surround sound experience.


These new techniques have the ability to isolate all of the spatial sonic cues in the stereo audio and then place each of these cues (segments) into the appropriate location of the surround field. This is done without any of the aforementioned unnatural artifacts. The result is an experience that appears rich, musical, natural, and as if the performance is occurring all around the listener. A simple description might be “sound all around.”


For example: The stereo mix of the song “Leaving On A Jet Plane,” as performed by Peter, Paul, and Mary, was produced with the three performers mixed in the following manner: Peter Yarrow in the left channel, Mary Travers in the center (i.e., her vocals are equally distributed between both left and right channels to create a phantom or pseudo center channel), and Paul Stookey in the right channel. When this song is heard using this new technique, Mary is heard in the center channel, Peter is heard in the left-rear channel, and Paul is heard in the right-rear channel. There is true isolation and distinction to the sound field, and is presented in a natural and musical fashion.


The techniques disclosed here may find particular application in the field of consumer audio. For example, the techniques disclosed herein may be applied to enhance any existing stereo sound device: radio, television, CD player, iPod/iPhone, wireless home sound system, car audio, and many more.


A particularly good candidate for these techniques may be the automobile listening experience. Adding this techniques to the in-car sound system would enable natural sounding surround from any stereo signal. Conveniently, many cars already have the needed number of speakers installed to enable the experience.


Another potential application relates to a peak limiter where the limiting function generates added harmonic spectra. The techniques disclosed herein may be used to perform the desired peak limiting, yet eliminate the added undesirable spectral information.





BRIEF DESCRIPTION OF THE DRAWINGS

The accompanying drawings, which are incorporated in and constitute a part of the specification, illustrate various example systems, methods, and so on, that illustrate various example embodiments of aspects of the invention. It will be appreciated that the illustrated element boundaries (e.g., boxes, groups of boxes, or other shapes) in the figures represent one example of the boundaries. One of ordinary skill in the art will appreciate that one element may be designed as multiple elements or that multiple elements may be designed as one element. An element shown as an internal component of another element may be implemented as an external component and vice versa. Furthermore, elements may not be drawn to scale.



FIG. 1 illustrates a block diagram of an exemplary extractor for extracting a common signal from at least two audio signals.



FIG. 2 illustrates a block diagram of an exemplary upmixer that may be applied to stereo audio signals to derive additional audio channels, from stereo to surround 5.1.



FIG. 3 illustrates a block diagram of an exemplary peak limiter cleanser.



FIG. 4 illustrates a flow diagram for an exemplary method for extracting a common signal from at least two audio signals.



FIG. 5 illustrates a block diagram of an exemplary machine for extracting a common signal from at least two audio signals.





DETAILED DESCRIPTION

Stereophonic audio offers two signal channels, left and right speaker channels. In stereo, audio cues are physically presented to the listener from the left and right side speakers. There is no physical center speaker. What appears as “centered” audio is actualized from an equal amount of the same content routed or panned to the left and right channels. This creates a phantom or pseudo center channel. Using a musical example, there may be a guitar located in the left channel, piano in the right channel, and an equal amount of a vocalist routed to both left and right channels. This configuration creates the illusion that the vocalist appears in the center, but there is no true center channel.


The techniques presented herein may receive the stereo left and right channel signals and yield sound stage cues which represent left, right, and center channels. A key part is isolating or extracting the center content. Once this is accomplished, the isolated center channel signal may be subtracted from the stereo left and right channel signals, and the result is isolated content for left, right, and center. The isolated left and right channel signals may respectively be used as rear-left and rear-right speaker signals. The front-left and front-right speaker signals may be a mixture of the original stereo cues attenuated by some amount. The extracted common signal may be routed to a physical center channel speaker. Low Frequency Extension (LFE) may be accomplished through the use of linearized filters designed to separate the bass frequencies. The resulting LFE speaker signal may be connected to a sub-woofer speaker. These six speaker signals yield the surround 5.1 sound stage.



FIG. 1 illustrates a block diagram of an exemplary extractor 10 for extracting a common signal from at least two audio signals. In the presence of multiple audio signals, which contain common and difference information, the extractor 10 extracts and isolates a common content signal C. In the example of FIG. 1 the extractor 10 extracts the common signal C from stereo left L and right R channel signals. The utility of the techniques disclosed herein to extract a common signal from at least two audio signals is not limited to the stereo example but include other context such as, as described in more detail below, eliminating undesired spectral information in peak limiting applications.


The extractor 10 includes a summer 12 that receives the L and R signals and sums them to obtain an L+R signal 13. The extractor 10 also includes a subtractor 14 that receives the L and R signals and subtract L from R to obtain an L−R signal 15. In the example of stereophonic (two channel) audio signals, common and difference information will be contained in the left and right channels. Adding and subtracting the two channels will yield sum and difference signals 13 and 15, thereby producing a monophonic mix of the stereo source L+R signal 13, and an isolation of the stereo sound field L−R signal 15, which represents content that resides in either the left or right channel.


The L+R and L-R signals 13 and 15 may then be converted from the time domain to the frequency domain through the use of the Fast Fourier Transform (FFT) or equivalent. To this end, the extractor 10 includes the FFT transformers 16 and 18, which transform the L+R and L−R signals 13 and 15 to their respective frequency domain representations 17 and 19.


The absolute value |X| of the frequency domain representation signals 17 and 19 is calculated, which yields the relative energy of each specified frequency bin within the FFT. To this end, the extractor 10 includes absolute value processors 20 and 22, which calculate and output the absolute value 21 of the frequency domain representation 17 of the L+R signal 13, and the absolute value 23 of the frequency domain representation 19 of the L−R signal 15.


The extractor 10 further includes subtractor 24 that subtracts the absolute value 23 of the frequency domain representation 19 of the L-R signal 15 from the absolute value 21 of the frequency domain representation 17 of the L+R signal 13 to obtain a difference signal 25. All uncommon detected spectra of the L+R and L-R signals is removed, allowing only detected spectra that is common to the original left and right channels to remain.


In the illustrated embodiment, the extractor 10 may also include gain adjustor or attenuator 26 which may adjust the gain of the difference signal 25. The gain adjustor 26 may adjust the gain of the difference signal 25 by, for example, −3.0 dB (0.707 attenuation).


This remaining detected spectra of the difference signal 25 (or of the gain adjusted difference signal 27) may then be applied to the frequency domain content 17 of the L+R signal. To this end, the extractor 10 may also include a multiplier 28 that multiplies the difference signal 25 or the gain adjusted difference signal 27 by the frequency domain representation 17 of the L+R signal 13 to obtain a product signal 29.


The extractor 10 may also include a divider 30 that divides the product signal 29 by the absolute value 21 of the frequency domain representation 17 of the L+R signal 13. In one embodiment, the divider 30 divides the product signal 29 by the absolute value 21 slightly adjusted upwards (e.g., +0.01, +0.001, or +0.0001) so that the absolute value 21 is higher than zero to avoid a division by zero. The divider 30 in essence normalizes the product signal 29 by dividing it by the absolute value 21 or by the absolute value 21 slightly adjusted upwards. The output 31 of the divider 30 corresponds to the frequency domain representation of the time domain common signal C.


The frequency domain representation 31 of the common signal C may then be passed through an Inverse Fast Fourier Transform (IFFT), which returns the common signal from the frequency domain back to the time domain. To this end, the extractor 10 includes a transformer 32 that transforms the frequency domain representation 31 to the common signal C.



FIG. 2 illustrates a block diagram of an exemplary upmixer 100 that may be applied to the stereo audio signals L, R to derive additional audio channels, from stereo to surround 5.1. The upmixer 100 incorporates the novel extractor 10 for extracting a common signal C from multiple audio signals L, R. Once this is accomplished, the common signal C may be isolated content for center, the common signal C may be subtracted from the stereo left and right channel signals, and the result may be isolated content for left and right. The isolated left and right channel signals may respectively be used as rear-left and rear-right speaker signals. The front-left and front-right speaker signals may be a mixture of the original stereo cues attenuated by some amount.


Low Frequency Extension (LFE) may be accomplished through the use of linearized filters designed to separate the bass frequencies. To this end, the upmixer 100 may include a summer 112 that sums the L and R signals to obtain an L+R signal 113. The upmixer 100 may also include a low-pass filter 114 that passes the bass frequencies signal 115 and a delay 116 that delays the signal 115 to account for any delays in the extractor 10. The resulting LFE channel signal 117 may be connected to a sub-woofer speaker 118.


The extracted common signal C (or a real portion or absolute value of the extracted common signal C) may be routed to a physical center channel speaker 119.


To accomplish left-rear and right-rear the upmixer 100 may include delays 120 and 121 that delay the L and R signals, respectively, to account for delays in the extractor 10 and output delayed L and R signals, 122 and 123. The upmixer 100 may also include a gain adjustor or attenuator 124 to which the common signal C (or a real portion or absolute value of the extracted common signal C) may be applied to obtain attenuated common signal 125. In one embodiment, the common signal C (or a real portion or absolute value of the extracted common signal C) may be attenuated by, for example, −3.0 dB (0.707 attenuation). In one embodiment, the common signal C (or a real portion or absolute value of the extracted common signal C) is not attenuated. The upmixer 100 may also include subtractors 126 and 127. The subtractor 126 subtracts the signal 125 from the delayed L signal 122 to output the left-rear signal 128. The subtractor 127 subtracts the signal 125 from the delayed R signal 123 to output the right-rear signal 129. The left-rear signal 128 may be routed to a left-rear speaker 130 and the right-rear signal may be routed to a right-rear speaker 131.


To accomplish left-front and right-front the upmixer 100 may include a gain adjustor or attenuator 132 to which the common signal C (or a real portion or absolute value of the extracted common signal C) may be applied to obtain attenuated common signal 133. In one embodiment, the common signal C (or a real portion or absolute value of the extracted common signal C) may be attenuated by, for example, −6 dB (0.5 attenuation). In one embodiment, the common signal C (or a real portion or absolute value of the extracted common signal C) is not attenuated. The upmixer 100 may also include summers 134 and 135. The summer 134 sums the signal 133 to the left-rear signal 128 to obtain the signal 136. The summer 135 sums the signal 133 to the right-rear signal 129 to obtain the signal 137. The upmixer 100 may also include gain adjustors or attenuators 138 and 139. The signal 136 may be applied to the attenuator 138 to obtain the left-front signal 140. The signal 137 may be applied to the attenuator 139 to obtain the right-front signal 141. In one embodiment, the signals 136 and 137 may be attenuated by, for example, −6 dB (0.5 attenuation). In one embodiment, the signals 136 and 137 are not attenuated. The left-front signal 140 may be routed to a left-front speaker 142 and the right-front signal may be routed to a right-front speaker 143.


The upmixer 100 isolates all of the spatial sonic cues in the stereo audio and then places these cues (segments) into the appropriate location of the surround field, significantly improving distinction in the sound field. Importantly, the techniques accomplish this without adding any fake or false signal cues, such as time delay, reverb, or ancillary unnatural effects to generate the surround sound experience. The result is an experience that appears rich, musical, natural, and as if the performance is occurring all around the listener. A simple description might be “sound all around.”



FIG. 3 illustrates a block diagram of an exemplary peak limiter cleanser 200. Peak limiters have become ubiquitous in music production. With the advent of digital processing, peak limiters can react instantaneously to or even in advance of audio peaks, ensuring that the signal never gets louder than a predetermined ceiling. Conventional peak limiters, however, come at the cost of clarity and punch, and some added spectral artifacts. Thus, conventional peak limiters may do a good job at limiting peaks but they add harmonic spectra that disturbs the sound. The peak limiter cleanser 200 addresses this problem.


The peak limiter cleanser 200 receives the original audio signal O and a peak limited version PL of the original audio signal. The signal PL has been peak limited using a conventional peak limiter. The peak limiter cleanser 200 incorporates the novel extractor 10 for extracting a common signal C from the original audio signal O and the peak limited version PL of the original audio signal to output a cleansed peak limited signal PL′. The signal PL′ represents a second peak-limited version of the original signal O but without at least some of the undesirable spectral information produced by conventional peak limiters. The signal PL′ provides significant improvements in clarity and punch because it does not include the undesirable spectral artifacts, thereby producing much more pleasant sound.


Example methods may be better appreciated with reference to flow diagrams.



FIG. 4 illustrates a flow diagram for an exemplary method 300 for extracting a common signal C from at least two audio signals, A and B. At 305, the method 300 includes receiving the two audio signals A and B. At 310, the method 300 includes summing the first signal A and the second signal B to obtain D. At 315, the method 300 includes subtracting the second signal B from the first signal A to obtain E.


At 320, the method 300 includes transforming the D signal to a frequency domain representation F. At 325, the method 300 includes transforming the E signal to a frequency domain representation G. At 330, the method 300 includes calculating absolute value of the frequency domain representation F to obtain H. At 335, the method 300 includes calculating absolute value of the frequency domain representation G to obtain I.


At 340, the method 300 includes subtracting the absolute value I from the absolute value H to obtain a difference signal J. At 345, the method 300 includes multiplying the difference signal J (or the difference signal J with its gain adjusted) by the frequency domain representation F to obtain a product signal K. At 350, the method 300 includes dividing the product signal K by the absolute value H (or by the absolute value H adjusted upwards) to obtain a frequency domain representation M of the common signal C.


At 355, the method 300 includes transforming the frequency domain representation M to the time domain common signal C.


While FIG. 4 illustrates various actions occurring in serial, it is to be appreciated that various actions illustrated could occur substantially in parallel, and while actions may be shown occurring in parallel, it is to be appreciated that these actions could occur substantially in series. While a number of processes are described in relation to the illustrated methods, it is to be appreciated that a greater or lesser number of processes could be employed and that lightweight processes, regular processes, threads, and other approaches could be employed. It is to be appreciated that other example methods may, in some cases, also include actions that occur substantially in parallel. The illustrated exemplary methods and other embodiments may operate in real-time, faster than real-time in a software or hardware or hybrid software/hardware implementation, or slower than real time in a software or hardware or hybrid software/hardware implementation.


While for purposes of simplicity of explanation, the illustrated methodologies are shown and described as a series of blocks, it is to be appreciated that the methodologies are not limited by the order of the blocks, as some blocks can occur in different orders or concurrently with other blocks from that shown and described. Moreover, less than all the illustrated blocks may be required to implement an example methodology. Furthermore, additional methodologies, alternative methodologies, or both can employ additional blocks, not illustrated.


In the flow diagram, blocks denote “processing blocks” that may be implemented with logic. The processing blocks may represent a method step or an apparatus element for performing the method step. The flow diagrams do not depict syntax for any particular programming language, methodology, or style (e.g., procedural, object-oriented). Rather, the flow diagram illustrates functional information one skilled in the art may employ to develop logic to perform the illustrated processing. It will be appreciated that in some examples, program elements like temporary variables, routine loops, and so on, are not shown. It will be further appreciated that electronic and software applications may involve dynamic and flexible processes so that the illustrated blocks can be performed in other sequences that are different from those shown or that blocks may be combined or separated into multiple components. It will be appreciated that the processes may be implemented using various programming approaches like machine language, procedural, object oriented or artificial intelligence techniques.



FIG. 5 illustrates a block diagram of an exemplary machine 400 for extracting a common signal from at least two audio signals. The machine 400 includes a processor 402, a memory 404, and I/O Ports 410 operably connected by a bus 408.


In one example, the machine 400 may transmit input and output signals including the signals L, R, C, O, PL, A, B, etc. described above via, for example, I/O Ports 410 or I/O Interfaces 418. The machine 400 may also include the extractor 10, upmixer 100, and peak limiter cleanser 200 and all of their components. Thus, the extractor 10, upmixer 100, and peak limiter cleanser 200 may be implemented in machine 400 as hardware, firmware, software, or combinations thereof and, thus, the machine 400 and its components may provide means for performing functions described herein as performed by the extractor 10, upmixer 100, and peak limiter cleanser 200.


The processor 402 can be a variety of various processors including dual microprocessor and other multi-processor architectures. The memory 404 can include volatile memory or non-volatile memory. The non-volatile memory can include, but is not limited to, ROM, PROM, EPROM, EEPROM, and the like. Volatile memory can include, for example, RAM, synchronous RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDR SDRAM), and direct RAM bus RAM (DRRAM).


A disk 406 may be operably connected to the machine 400 via, for example, an I/O Interfaces (e.g., card, device) 418 and an I/O Ports 410. The disk 406 can include, but is not limited to, devices like a magnetic disk drive, a solid state disk drive, a floppy disk drive, a tape drive, a Zip drive, a flash memory card, or a memory stick. Furthermore, the disk 406 can include optical drives like a CD-ROM, a CD recordable drive (CD-R drive), a CD rewriteable drive (CD-RW drive), or a digital video ROM drive (DVD ROM). The memory 404 can store processes 414 or data 416, for example. The disk 406 or memory 404 can store an operating system that controls and allocates resources of the machine 400.


The bus 408 can be a single internal bus interconnect architecture or other bus or mesh architectures. While a single bus is illustrated, it is to be appreciated that machine 400 may communicate with various devices, logics, and peripherals using other busses that are not illustrated (e.g., PCIE, SATA, Infiniband, 1394, USB, Ethernet). The bus 408 can be of a variety of types including, but not limited to, a memory bus or memory controller, a peripheral bus or external bus, a crossbar switch, or a local bus. The local bus can be of varieties including, but not limited to, an industrial standard architecture (ISA) bus, a microchannel architecture (MCA) bus, an extended ISA (EISA) bus, a peripheral component interconnect (PCI) bus, a universal serial (USB) bus, and a small computer systems interface (SCSI) bus.


The machine 400 may interact with input/output devices via I/O Interfaces 418 and I/O Ports 410. Input/output devices can include, but are not limited to, a keyboard, a microphone, a pointing and selection device, cameras, video cards, displays, disk 406, network devices 420, and the like. The I/O Ports 410 can include but are not limited to, serial ports, parallel ports, and USB ports.


The machine 400 can operate in a network environment and thus may be connected to network devices 420 via the I/O Interfaces 418, or the I/O Ports 410. Through the network devices 420, the machine 400 may interact with a network. Through the network, the machine 400 may be logically connected to remote devices. The networks with which the machine 400 may interact include, but are not limited to, a local area network (LAN), a wide area network (WAN), and other networks. The network devices 420 can connect to LAN technologies including, but not limited to, fiber distributed data interface (FDDI), copper distributed data interface (CDDI), Ethernet (IEEE 802.3), token ring (IEEE 802.5), wireless computer communication (IEEE 802.11), Bluetooth (IEEE 802.15.1), Zigbee (IEEE 802.15.4) and the like. Similarly, the network devices 420 can connect to WAN technologies including, but not limited to, point to point links, circuit switching networks like integrated services digital networks (ISDN), packet switching networks, and digital subscriber lines (DSL). While individual network types are described, it is to be appreciated that communications via, over, or through a network may include combinations and mixtures of communications.


While example systems, methods, and so on, have been illustrated by describing examples, and while the examples have been described in considerable detail, it is not the intention of the applicants to restrict or in any way limit scope to such detail. It is, of course, not possible to describe every conceivable combination of components or methodologies for purposes of describing the systems, methods, and so on, described herein. Additional advantages and modifications will readily appear to those skilled in the art. Therefore, the invention is not limited to the specific details, the representative apparatus, and illustrative examples shown and described. Thus, this application is intended to embrace alterations, modifications, and variations that fall within the scope of the appended claims. Furthermore, the preceding description is not meant to limit the scope of the invention. Rather, the scope of the invention is to be determined by the appended claims and their equivalents.


To the extent that the term “includes” or “including” is employed in the detailed description or the claims, it is intended to be inclusive in a manner similar to the term “comprising” as that term is interpreted when employed as a transitional word in a claim. Furthermore, to the extent that the term “or” is employed in the detailed description or claims (e.g., A or B) it is intended to mean “A or B or both”. When the applicants intend to indicate “only A or B but not both” then the term “only A or B but not both” will be employed. Thus, use of the term “or” herein is the inclusive, and not the exclusive use. See, Bryan A. Garner, A Dictionary of Modern Legal Usage 624 (2d. Ed. 1995).

Claims
  • 1. A machine or group of machines for extracting a common signal from at least two audio signals, comprising: a summer configured to receive the at least two audio signals including a first signal and a second signal and to sum the first signal and the second signal to obtain a first+second signal;a subtractor configured to subtract the second signal from the first signal to obtain a first−second signal;at least one transformer configured to transform the first+second signal and the first−second signal to frequency domain representations of the first+second signal and the first−second signal, respectively;a processor configured to calculate absolute value of the frequency domain representation of the first+second signal and absolute value of the frequency domain representation of the first−second signal;a subtractor configured to subtracting the absolute value of the frequency domain representation of the first−second signal from the absolute value of the frequency domain representation of the first+second signal to obtain a difference signal;a multiplier configured to multiply the difference signal or the difference signal with its gain adjusted times the frequency domain representation of the first+second signal to obtain a product signal;a divider configured to divide the product signal by the absolute value of the frequency domain representation of the first+second signal or by the absolute value of the frequency domain representation of the first+second signal adjusted upwards, to obtain a frequency domain representation of the common signal; anda transformer configured to transform the frequency domain representation of the common signal to the common signal in the time domain.
  • 2. The machine or group of machines of claim 1, wherein the at least two audio signals correspond to a left channel and a right channel in a stereo system.
  • 3. The machine or group of machine of claim 2, comprising: a subtractor configured to subtract the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted from the first signal or the first signal delayed to obtain a left rear signal or a right rear signal in a surround system.
  • 4. The machine or group of machines of claim 2, comprising: a subtractor configured to subtract the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted from the first signal or the first signal delayed to obtain a left rear signal or a right rear signal; anda summer configured to sum the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted to the left rear signal or the right rear signal to obtain a sum, andan attenuator configured to adjust gain of the sum to obtain a left front signal or a right front signal in a surround system.
  • 5. The machine or group of machines of claim 2, comprising: a loudspeaker configured to receive a portion of the common signal as a center channel signal.
  • 6. The machine or group of machines of claim 2, comprising: a low-pass filter and a delayer configured to filter and delay the first+second signal to obtain a low frequency effect signal.
  • 7. The machine or group of machines of claim 1, wherein the at least two audio signals correspond to an original signal and a first peak-limited version of the original signal, the common signal representing a second peak-limited version of the original signal without at least some undesirable spectral information.
  • 8. A method for extracting a common signal from at least two audio signals, the method comprising: receiving the at least two audio signals including a first signal and a second signal;summing the first signal and the second signal to obtain a first+second signal;subtracting the second signal from the first signal to obtain a first−second signal;transforming the first+second signal and the first−second signal to frequency domain representations of the first+second signal and the first−second signal, respectively;calculating absolute value of the frequency domain representation of the first+second signal and absolute value of the frequency domain representation of the first−second signal;subtracting the absolute value of the frequency domain representation of the first−second signal from the absolute value of the frequency domain representation of the first+second signal to obtain a difference signal;multiplying the difference signal or the difference signal with its gain adjusted times the frequency domain representation of the first+second signal to obtain a product signal;dividing the product signal by the absolute value of the frequency domain representation of the first+second signal or by the absolute value of the frequency domain representation of the first+second signal adjusted upwards, to obtain a frequency domain representation of the common signal; andtransforming the frequency domain representation of the common signal to the common signal in the time domain.
  • 9. The method of claim 8, wherein the at least two audio signals correspond to a left channel and a right channel in a stereo system.
  • 10. The method of claim 9, comprising: subtracting the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted from the first signal or the first signal delayed to obtain a left rear signal or a right rear signal in a surround system.
  • 11. The method of claim 9, comprising: subtracting the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted from the first signal or the first signal delayed to obtain a left rear signal or a right rear signal; andsumming the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted to the left rear signal or the right rear signal, and adjusting gain of the sum to obtain a left front signal or a right front signal in a surround system.
  • 12. The method of claim 9, comprising: applying a portion of the common signal as a center channel signal.
  • 13. The method of claim 9, comprising: low-pass filtering and delaying the first+second signal to obtain a low frequency effect signal.
  • 14. The method of claim 8, wherein the at least two audio signals correspond to an original signal and a first peak-limited version of the original signal, the common signal representing a second peak-limited version of the original signal without at least some undesirable spectral information.
  • 15. A computer program product for extracting a common signal from at least two audio signals, the computer program product includes a non-transient computer-readable medium and comprising machine executable instructions which, when executed, cause the machine to perform steps of a method comprising: receiving the at least two audio signals including a first signal and a second signal;summing the first signal and the second signal to obtain a first+second signal;subtracting the second signal from the first signal to obtain a first−second signal;transforming the first+second signal and the first−second signal to frequency domain representations of the first+second signal and the first−second signal, respectively;calculating absolute value of the frequency domain representation of the first+second signal and absolute value of the frequency domain representation of the first−second signal;subtracting the absolute value of the frequency domain representation of the first−second signal from the absolute value of the frequency domain representation of the first+second signal to obtain a difference signal;multiplying the difference signal or the difference signal with its gain adjusted times the frequency domain representation of the first+second signal to obtain a product signal;dividing the product signal by the absolute value of the frequency domain representation of the first+second signal or by the absolute value of the frequency domain representation of the first+second signal adjusted upwards, to obtain a frequency domain representation of the common signal; andtransforming the frequency domain representation of the common signal to the common signal in the time domain.
  • 16. The computer program product of claim 15, wherein the at least two audio signals correspond to a left channel and a right channel in a stereo system.
  • 17. The computer program product of claim 16, comprising: subtracting the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted from the first signal or the first signal delayed to obtain a left rear signal or a right rear signal in a surround system.
  • 18. The computer program product of claim 16, comprising: subtracting the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted from the first signal or the first signal delayed to obtain a left rear signal or a right rear signal; andsumming the common signal, a portion of the common signal, the common signal with its gain adjusted or the portion with its gain adjusted to the left rear signal or the right rear signal, and adjusting gain of the sum to obtain a left front signal or a right front signal in a surround system.
  • 19. The computer program product of claim 16, comprising: low-pass filtering and delaying the first+second signal to obtain a low frequency effect signal.
  • 20. The computer program product of claim 15, wherein the at least two audio signals correspond to an original signal and a first peak-limited version of the original signal, the common signal representing a second peak-limited version of the original signal without at least some undesirable spectral information.
US Referenced Citations (372)
Number Name Date Kind
2655564 Aage Oct 1953 A
3070669 Franssen et al. Dec 1962 A
3272906 De Vries et al. Sep 1966 A
3987256 Nagamura Oct 1976 A
4027101 DeFreitas et al. May 1977 A
4039755 Berkovitz Aug 1977 A
4048442 Mannila et al. Sep 1977 A
4049912 Mitchell Sep 1977 A
4053711 DeFreitas et al. Oct 1977 A
4105864 Berkovitz Aug 1978 A
4251688 Furner Feb 1981 A
4332979 Fischer Jun 1982 A
4524451 Watanabe Jun 1985 A
4574391 Morishima Mar 1986 A
4594729 Weingartner Jun 1986 A
4856064 Iwamatsu Aug 1989 A
4941177 Mandell et al. Jul 1990 A
5068897 Yamato et al. Nov 1991 A
5119422 Price Jun 1992 A
5121433 Kendall et al. Jun 1992 A
5193118 Latham-Brown et al. Mar 1993 A
5235646 Wilde et al. Aug 1993 A
5255370 Sako et al. Oct 1993 A
5261005 Masayuki Nov 1993 A
5263087 Fosgate Nov 1993 A
5280528 Fosgate Jan 1994 A
5386473 Harrison Jan 1995 A
5414774 Yumoto May 1995 A
5428687 Willcocks et al. Jun 1995 A
5594800 Gerzon Jan 1997 A
5644640 Fosgate Jul 1997 A
5666424 Fosgate et al. Sep 1997 A
5796844 Griesinger Aug 1998 A
5812674 Jot et al. Sep 1998 A
5854847 Yoshida Dec 1998 A
6072878 Moorer Jun 2000 A
6252965 Beard Jun 2001 B1
6405163 LaRoche Jun 2002 B1
6684060 Curtin Jan 2004 B1
6804565 Eid Oct 2004 B2
6829018 Lin et al. Dec 2004 B2
6836295 Cooper Dec 2004 B1
6920233 Wolfson et al. Jul 2005 B2
7243073 Yeh et al. Jul 2007 B2
7257231 Avendano et al. Aug 2007 B1
7292901 Baumgarte et al. Nov 2007 B2
7330552 Lamance Feb 2008 B1
7353169 Goodwin Apr 2008 B1
7394903 Herre et al. Jul 2008 B2
7412380 Avendano et al. Aug 2008 B1
7542815 Berchin et al. Jun 2009 B1
7567845 Avendano et al. Jul 2009 B1
7606373 Moorer Oct 2009 B2
7644003 Baumgarte et al. Jan 2010 B2
7650000 Kawana et al. Jan 2010 B2
7693721 Baumgarte et al. Apr 2010 B2
7710499 Cooper May 2010 B2
7760890 Furge et al. Jul 2010 B2
7769178 Beard Aug 2010 B2
7769179 Beard Aug 2010 B2
7769180 Beard Aug 2010 B2
7769181 Beard Aug 2010 B2
7773756 Beard Aug 2010 B2
7773757 Beard Aug 2010 B2
7773758 Beard Aug 2010 B2
7783052 Beard Aug 2010 B2
7792304 Beard Sep 2010 B2
7792305 Beard Sep 2010 B2
7792306 Beard Sep 2010 B2
7792307 Beard Sep 2010 B2
7792308 Beard Sep 2010 B2
7796765 Beard Sep 2010 B2
7856110 Fujita et al. Dec 2010 B2
7864964 Beard Jan 2011 B2
7864965 Beard Jan 2011 B2
7864966 Beard Jan 2011 B2
7873171 Beard Jan 2011 B2
7876905 Beard Jan 2011 B2
7965849 Beard Jun 2011 B2
7983424 Kjorling et al. Jul 2011 B2
7986796 Kim et al. Jul 2011 B2
8014535 Beard Sep 2011 B2
8015368 Sharma et al. Sep 2011 B2
8019093 Avendano et al. Sep 2011 B2
8027480 Beard Sep 2011 B2
8031879 Eid et al. Oct 2011 B2
8041043 Faller Oct 2011 B2
8054980 Wu et al. Nov 2011 B2
8099293 Kulkarni Jan 2012 B2
8108220 Saunders et al. Jan 2012 B2
8116479 Hamada et al. Feb 2012 B2
8126172 Horbach et al. Feb 2012 B2
8159610 Cooper Apr 2012 B2
8194860 Sugawara Jun 2012 B2
8200500 Baumgarte et al. Jun 2012 B2
8204237 Goodwin Jun 2012 B2
8280077 Avendano et al. Oct 2012 B2
8280743 Seefeldt et al. Oct 2012 B2
8290167 Pulkki et al. Oct 2012 B2
8295496 Kulkarni et al. Oct 2012 B2
8300833 Beard Oct 2012 B2
8379868 Goodwin et al. Feb 2013 B2
8472638 Furge et al. Jun 2013 B2
8520862 Scholz Aug 2013 B2
8589634 Sharma et al. Nov 2013 B2
8611550 Del Galdo et al. Dec 2013 B2
8619998 Walsh et al. Dec 2013 B2
8670850 Soulodre Mar 2014 B2
8675881 Hultz et al. Mar 2014 B2
8705779 Lee Apr 2014 B2
8731209 Uhle et al. May 2014 B2
8744247 Crockett et al. Jun 2014 B2
8767969 Laroche et al. Jul 2014 B1
8774417 Cabot et al. Jul 2014 B1
8787585 Brown Jul 2014 B2
8831931 Kuntz et al. Sep 2014 B2
8908873 Herre et al. Dec 2014 B2
9055381 Kirkeby et al. Jun 2015 B2
9071723 Cooper Jun 2015 B2
9082395 Heiko et al. Jul 2015 B2
9084068 Abrahamsson et al. Jul 2015 B2
9094771 Tsingos et al. Jul 2015 B2
9100767 Van Baelen et al. Aug 2015 B2
9161149 Visser et al. Oct 2015 B2
9179236 Robinson et al. Nov 2015 B2
9183839 Faller Nov 2015 B2
9233308 Kuo et al. Jan 2016 B2
9271102 Kulavik et al. Feb 2016 B2
9344825 Carroll May 2016 B2
9357325 Lee et al. May 2016 B2
9368122 Kuntz et al. Jun 2016 B2
9369164 Kim et al. Jun 2016 B2
9408010 Harma et al. Aug 2016 B2
9443524 Wu et al. Sep 2016 B2
9462384 Lakkundi et al. Oct 2016 B2
9462405 Lee Oct 2016 B2
9478228 Oomen et al. Oct 2016 B2
9485601 Cabot et al. Nov 2016 B1
9503816 Iseki et al. Nov 2016 B2
9549276 Chon et al. Jan 2017 B2
9578437 Oh et al. Feb 2017 B2
9583113 Kapinos Feb 2017 B2
9628934 Davis Apr 2017 B2
20030039365 Eid et al. Feb 2003 A1
20030039366 Eid et al. Feb 2003 A1
20030040822 Eid et al. Feb 2003 A1
20030223589 Wang Dec 2003 A1
20030235317 Baumgarte Dec 2003 A1
20030236583 Baumgarte et al. Dec 2003 A1
20040008847 Kim et al. Jan 2004 A1
20040062401 Davis Apr 2004 A1
20040212320 Dowling et al. Oct 2004 A1
20050012860 Cooper Jan 2005 A1
20050018860 Furge et al. Jan 2005 A1
20050063551 Cheng et al. Mar 2005 A1
20050141724 Hesdahl Jun 2005 A1
20050141728 Moorer Jun 2005 A1
20050152556 Brett Jul 2005 A1
20050165502 Hanes Jul 2005 A1
20050271215 Kulkarni Dec 2005 A1
20050276420 Davis Dec 2005 A1
20060088168 Beard Apr 2006 A1
20060088175 Eid et al. Apr 2006 A1
20060115100 Faller Jun 2006 A1
20060198528 Tuffy et al. Sep 2006 A1
20060256969 Takashima et al. Nov 2006 A1
20060262939 Buchner et al. Nov 2006 A1
20070047736 Shioda Mar 2007 A1
20070135952 Chubarev Jun 2007 A1
20070140497 Moon et al. Jun 2007 A1
20070140498 Moon et al. Jun 2007 A1
20070183601 Van Loon et al. Aug 2007 A1
20070206800 Beard Sep 2007 A1
20070206801 Beard Sep 2007 A1
20070206802 Beard Sep 2007 A1
20070206803 Beard Sep 2007 A1
20070206804 Beard Sep 2007 A1
20070206805 Beard Sep 2007 A1
20070206806 Beard Sep 2007 A1
20070206807 Beard Sep 2007 A1
20070206808 Beard Sep 2007 A1
20070206809 Beard Sep 2007 A1
20070206810 Beard Sep 2007 A1
20070206811 Beard Sep 2007 A1
20070206812 Beard Sep 2007 A1
20070206813 Beard Sep 2007 A1
20070206814 Beard Sep 2007 A1
20070206815 Beard Sep 2007 A1
20070206816 Beard Sep 2007 A1
20070206821 Beard Sep 2007 A1
20070211905 Beard Sep 2007 A1
20070233293 Villemoes et al. Oct 2007 A1
20070263877 Beard Nov 2007 A1
20070269063 Goodwin et al. Nov 2007 A1
20080025534 Kuhn et al. Jan 2008 A1
20080031462 Walsh et al. Feb 2008 A1
20080059160 Saunders et al. Mar 2008 A1
20080069366 Soulodre Mar 2008 A1
20080091436 Breebaart et al. Apr 2008 A1
20080091439 Baumgarte et al. Apr 2008 A1
20080152155 Avendano et al. Jun 2008 A1
20080159571 Hooley Jul 2008 A1
20080175394 Goodwin Jul 2008 A1
20080232603 Soulodre et al. Sep 2008 A1
20080232616 Pulkki et al. Sep 2008 A1
20080263285 Sharma et al. Oct 2008 A1
20080294444 Oh et al. Nov 2008 A1
20080298612 Kulkarni Dec 2008 A1
20080304671 Kulkarni Dec 2008 A1
20080317257 Furge et al. Dec 2008 A1
20080319564 Furge et al. Dec 2008 A1
20090092259 Jot et al. Apr 2009 A1
20090110204 Walsh et al. Apr 2009 A1
20090129603 Cho May 2009 A1
20090180625 Bai et al. Jul 2009 A1
20090198356 Goodwin et al. Aug 2009 A1
20090225991 Oh et al. Sep 2009 A1
20090287328 Berchin Nov 2009 A1
20090319281 Baumgarte et al. Dec 2009 A1
20100030563 Uhle et al. Feb 2010 A1
20100061558 Faller Mar 2010 A1
20100215199 Breebaart Aug 2010 A1
20100246833 Hagiwara et al. Sep 2010 A1
20100296672 Vickers Nov 2010 A1
20110002488 Van Daele et al. Jan 2011 A1
20110026720 Ohta Feb 2011 A1
20110081024 Soulodre Apr 2011 A1
20110129091 Kron Jun 2011 A1
20110137662 McGrath et al. Jun 2011 A1
20110150228 Yoon et al. Jun 2011 A1
20110164855 Crockett et al. Jul 2011 A1
20110200195 Lau et al. Aug 2011 A1
20120016502 Sharma et al. Jan 2012 A1
20120039477 Schijers et al. Feb 2012 A1
20120045065 Ishihara Feb 2012 A1
20120059498 Berchin Mar 2012 A1
20120063608 Soulodre Mar 2012 A1
20120070006 Berchin Mar 2012 A1
20120099739 Hultz et al. Apr 2012 A1
20120127365 Cooper May 2012 A1
20120128160 Kim et al. May 2012 A1
20120155650 Horbach Jun 2012 A1
20120195434 Herre et al. Aug 2012 A1
20120207309 Eppolito Aug 2012 A1
20120266067 Armstrong et al. Oct 2012 A1
20120275613 Soulodre Nov 2012 A1
20120300946 Ma Nov 2012 A1
20120307048 Abrahamsson et al. Dec 2012 A1
20120328109 Harma et al. Dec 2012 A1
20120328136 Chiang Dec 2012 A1
20130021502 Oku et al. Jan 2013 A1
20130070927 Harma Mar 2013 A1
20130108057 Yuyama et al. May 2013 A1
20130142338 Chang et al. Jun 2013 A1
20130144922 Par Jun 2013 A1
20130173273 Kuntz et al. Jul 2013 A1
20130208895 Horbach Aug 2013 A1
20130251156 Katayama Sep 2013 A1
20130315402 Visser et al. Nov 2013 A1
20130322633 Stone et al. Dec 2013 A1
20140010375 Usher et al. Jan 2014 A1
20140023094 Chen Jan 2014 A1
20140072124 Stoecklmeier et al. Mar 2014 A1
20140072126 Uhle et al. Mar 2014 A1
20140086437 Koh et al. Mar 2014 A1
20140119545 Uhle et al. May 2014 A1
20140153744 Brannmark et al. Jun 2014 A1
20140185812 Van Achte et al. Jul 2014 A1
20140205100 Faller et al. Jul 2014 A1
20140233762 Vilkamo et al. Aug 2014 A1
20140270182 Vilermo et al. Sep 2014 A1
20140270281 Walsh et al. Sep 2014 A1
20140270721 Crockett et al. Sep 2014 A1
20140282706 Kim et al. Sep 2014 A1
20140336800 Radhakrishnan et al. Nov 2014 A1
20140341404 Choisel et al. Nov 2014 A1
20140355769 Peters et al. Dec 2014 A1
20140355770 Peters et al. Dec 2014 A1
20140355771 Peters et al. Dec 2014 A1
20140358266 Peters et al. Dec 2014 A1
20140358558 Sen et al. Dec 2014 A1
20140358559 Sen et al. Dec 2014 A1
20140358560 Sen et al. Dec 2014 A1
20140358561 Sen et al. Dec 2014 A1
20140358562 Sen et al. Dec 2014 A1
20140358563 Sen et al. Dec 2014 A1
20140358564 Sen et al. Dec 2014 A1
20140358565 Peters et al. Dec 2014 A1
20140358567 Koppens et al. Dec 2014 A1
20140373044 Carrol et al. Dec 2014 A1
20150003616 Middlemiss et al. Jan 2015 A1
20150036829 Meier Feb 2015 A1
20150036849 Thompson Feb 2015 A1
20150042880 Cooper Feb 2015 A1
20150063610 Mossner Mar 2015 A1
20150092965 Umminger et al. Apr 2015 A1
20150098597 Kulavik Apr 2015 A1
20150106475 Tan et al. Apr 2015 A1
20150110275 Tammi et al. Apr 2015 A1
20150131826 Bahne et al. May 2015 A1
20150156599 Romigh Jun 2015 A1
20150172843 Quan Jun 2015 A1
20150179180 Oh et al. Jun 2015 A1
20150213807 Breebaart et al. Jul 2015 A1
20150215721 Sato et al. Jul 2015 A1
20150243289 Radhakrishnan et al. Aug 2015 A1
20150249899 Kuech et al. Sep 2015 A1
20150264502 Aoki et al. Sep 2015 A1
20150269948 Purnhagen et al. Sep 2015 A1
20150271620 Lando et al. Sep 2015 A1
20150277847 Yliaho et al. Oct 2015 A1
20150312692 Esslinger Oct 2015 A1
20150326988 Zielinsky Nov 2015 A1
20150332681 Kim et al. Nov 2015 A1
20150334500 Mieth et al. Nov 2015 A1
20150340043 Koppens et al. Nov 2015 A1
20150350801 Koppens et al. Dec 2015 A1
20150350802 Jo et al. Dec 2015 A1
20150350804 Crockett et al. Dec 2015 A1
20150356976 Herre et al. Dec 2015 A1
20150358754 Koppens et al. Dec 2015 A1
20150358756 Harma et al. Dec 2015 A1
20150371644 Par Dec 2015 A1
20150382128 Ridihalgh et al. Dec 2015 A1
20160005408 Visser et al. Jan 2016 A1
20160007133 Mateos et al. Jan 2016 A1
20160021476 Robinson et al. Jan 2016 A1
20160021477 Hiipakka et al. Jan 2016 A1
20160037283 Uhle et al. Feb 2016 A1
20160073213 Yamamoto et al. Mar 2016 A1
20160073215 De Bruijn et al. Mar 2016 A1
20160080883 Yamamoto et al. Mar 2016 A1
20160080884 Song et al. Mar 2016 A1
20160080886 De Bruijn et al. Mar 2016 A1
20160100253 Yliaho et al. Apr 2016 A1
20160100270 Araki Apr 2016 A1
20160112819 Mehnert et al. Apr 2016 A1
20160127847 Shi et al. May 2016 A1
20160127849 Suyama et al. May 2016 A1
20160133260 Hatanaka et al. May 2016 A1
20160133261 Shi et al. May 2016 A1
20160133263 Borss et al. May 2016 A1
20160142850 Borss et al. May 2016 A1
20160150343 Wang et al. May 2016 A1
20160157040 Ertel et al. Jun 2016 A1
20160165374 Shi et al. Jun 2016 A1
20160192102 Estrada et al. Jun 2016 A1
20160225377 Miyasaka et al. Aug 2016 A1
20160241982 Seefeldt et al. Aug 2016 A1
20160247514 Villemoes et al. Aug 2016 A1
20160255454 McGrath et al. Sep 2016 A1
20160269847 Par Sep 2016 A1
20160286331 Murayama et al. Sep 2016 A1
20160302021 Kim et al. Oct 2016 A1
20160309276 Ridihalgh et al. Oct 2016 A1
20160316109 Cooper Oct 2016 A1
20160330560 Chon et al. Nov 2016 A1
20160336014 Brockmole et al. Nov 2016 A1
20160366530 Peters et al. Dec 2016 A1
20160381482 Peters et al. Dec 2016 A1
20160381483 Robinson et al. Dec 2016 A1
20170024185 Lakkundi et al. Jan 2017 A1
20170026751 Suzuki et al. Jan 2017 A1
20170026768 Kron Jan 2017 A1
20170055096 Chamness et al. Feb 2017 A1
20170064484 Borss et al. Mar 2017 A1
20170070838 Helwani et al. Mar 2017 A1
20170078818 Habets et al. Mar 2017 A1
20170078819 Habets et al. Mar 2017 A1
20170094438 Chon et al. Mar 2017 A1
20170099546 Spitznagle et al. Apr 2017 A1
20170099558 Spitznagle et al. Apr 2017 A1
Foreign Referenced Citations (2)
Number Date Country
2012163445 Dec 2012 WO
2016130500 Aug 2016 WO