A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.
The listening environment, including speakers, room geometries and materials, furniture, and so forth can have an enormous effect on the quality of audio reproduction. Recently it has been shown that one can employ relatively simple digital filtering to provide a much more faithful reproduction of audio as it was originally recorded in a studio or concert hall (see, e.g., http://www.princeton.edu/3D3A/BACCH_intro.html). In fact, it is possible to produce three-dimensional sound using two speakers by using active cross-talk cancellation. In virtually any kind of listening environment, one can also compensate for speaker mismatches, and variability in the room arrangement, using phase and amplitude equalization. Today, however, with music being highly portable with mp3 players, mobile phones, and the like, and with music available through Internet cloud services, consumers bring their music into many different listening environments. It is rare that these environments are configured in an optimal way, and so it is advantageous to have a simple but effective method of calibrating digital filters for use with portable devices such as mobile phones, that can be used with various kinds of audio playback devices, such as automobile audio systems, phone docking systems, Internet connected speaker systems, and the like. In addition, audio that is played on laptops, TVs, tablets, etc. can also benefit from precise digital equalization. Systems and methods are presented herein for facilitating cost-effective calibration of filters for, e.g., correcting room and/or speaker-based distortion and/or binaural imbalances in audio reproduction, and/or for producing three-dimensional (3D) sound in stereo system environments.
The inventive body of work will be readily understood by referring to the following detailed description in conjunction with the accompanying drawings, in which:
A detailed description of the inventive body of work is provided below. While several embodiments are described, it should be understood that the inventive body of work is not limited to any one embodiment, but instead encompasses numerous alternatives, modifications, and equivalents. In addition, while numerous specific details are set forth in the following description in order to provide a thorough understanding of the inventive body of work, some embodiments can be practiced without some or all of these details. Moreover, for the purpose of clarity, certain technical material that is known in the related art has not been described in detail in order to avoid unnecessarily obscuring the inventive body work.
Embodiments of the disclosure may be understood by reference to the drawings, wherein like parts may be designated by like numerals. The components of the disclosed embodiments, as generally described and illustrated in the figures herein, could be arranged and designed in a wide variety of different configurations. Thus, the following detailed description of various embodiments is not intended to limit the scope of the disclosure, as claimed, but is merely representative of possible embodiments. In addition, the actions in the methods disclosed herein do not necessarily need to be performed in any specific order, or even sequentially, nor need the actions be performed only once, unless otherwise specified.
Systems and methods are presented for facilitating cost-effective calibration of filters for, e.g., correcting room and/or speaker-based distortion and/or binaural imbalances in audio reproduction, and/or for producing three-dimensional sound in stereo system environments.
Heretofore, calibration methods for filters have been cumbersome, inconvenient, and expensive, and are not easily performed by the user of an audio source in different environments. Some embodiments of the systems and methods described herein can be used by consumers without extensive knowledge or experience, using devices that the consumers already own and know how to use. Participation by the user should preferably take a relatively short amount of time (e.g., a few seconds or minutes). This will help facilitate more widespread performance of automatic equalization methods for many more audio sources in many more environments.
Systems and methods are described herein for addressing some or all of the following illustrative situations:
It will be appreciated that the examples in the foregoing list are provided for purposes of illustration and not limitation, and that embodiments of the systems and methods described herein could be applied in many other situations as well.
In one embodiment, device 104 is configured to send a predefined test file to the audio source device 101 (e.g., an Internet music repository, home network server, etc.) or otherwise causes the audio source device 101 to initiate playing of the requisite test file over one or more of speakers S1-Sn. In other embodiments, device 104 simply detects the playing of the file or other content via microphone 105. Upon receipt of the played back test file or other audio content via microphone 105, portable device (and/or a service or device in communication therewith) analyzes it in comparison to the original audio content and determines how to appropriately process future audio playback using DSP 102 and/or other means to improve the perceived quality of audio content to the recipient/user.
To improve performance, such analysis and processing may take into account the transfer function of the microphone 105 (which, as shown in
It will be appreciated that the system shown in
The device in this example may, e.g., be a mobile phone, tablet, laptop, or any other device that has a microphone and/or accommodates connection to a microphone. When the user runs the app, the app provides, e.g., through the user interface of the device, instructions for positioning the microphone to collect audio test data (202). For example, in one embodiment the app might instruct the user to position the microphone of the device next to his or her left ear and press a button (or other user input) on the device and to wait until an audio test file starts playing through one or more of the speakers S1 through Sn and then stops (203). In one embodiment, the app can control what audio test file to play. The user could then be instructed to reposition the microphone (204), e.g., by placing the microphone next to his or her right ear, at which point another (or the same) test file is played (205). Depending on the number of speakers in the system and/or the number of calibration tests, the user may be prompted to repeat this procedure a few times (e.g., a “yes” exit from block 206).
In one embodiment, with each test, a test result file is created or updated. For each test source, there will be an ideal test response. The device (or another system in communication therewith) will be able to calculate equalization parameters for each speaker in the system by performing spectral analysis on the received signal and comparing the ideal test response with the actual test response. For example, if the test source were an impulse function, the ideal response would have a flat frequency spectrum and the actual response would be easy to compare. However, for a number of reasons, different signals, selected to accommodate phase equalization and to deal with other types of impairments, may be used.
In one embodiment, calculation of the optimal equalization parameters is performed in a way that accommodates the transfer function of the microphone. This function will typically vary among different microphone designs, and so it will typically be important to have this information so that this transfer function can be subtracted out of the system. Thus, in some embodiments, a database (e.g., an Internet accessible database) of microphone transfer functions is maintained that can be referenced by the app. In the present case of the mobile smartphone, lookup of the transfer function is straightforward and can typically be performed by the app without any input from the user, because the app can reference the system information file of the smartphone to determine the model number of the phone, which can then be used to look up the transfer function in the database (106). The response curve may, for example, contain data such as illustrated at http://blog.faberacoustical.com/2009/ios/iphone/iphone-microphone-frequency-response-comparison, and this data can then be used in the computation of the optimal filter characteristics, as indicated above. In other embodiments, one or more transfer functions could be stored locally on the device itself, and no network connection would be needed.
Referring once again to
It will be appreciated that there are a number of variations of the systems and methods described herein for facilitating use of a portable device to calibrate digital filters that can optimize the function of speakers in a particular environment. For example, one way of simplifying the method described in connection with
It will also be appreciated that while certain examples have been described for facilitating calibration and optimization of speaker systems, some of the principles described herein are suitable for broader application. For example, without limitation, a device (e.g., a mobile phone, tablet, etc.) comprising a microphone and a speaker could be used to perform some or all of the following actions using audio detection and processing techniques such as those described above:
Using the ring tone as a probe signal.
Measuring room size.
Measuring the distance to another device.
Recognizing familiar locations by room response.
Detecting room features, like double-pane windows, narrow passages, and/or the like.
Mapping a room acoustically.
Detecting being outdoors.
Measuring temperature acoustically.
Identifying the bearer by voice (e.g., for detecting theft and/or positively identifying the user to facilitate device-sharing).
Detecting being submerged underwater.
Correlating acoustic data with camera data, GPS, etc.
Acoustic scene analysis (e.g., identification of other ring tones, ambient noises, sirens, alarms, familiar voices and sounds, etc.).
As shown in
One of ordinary skill in the art will appreciate that the systems and methods described herein can be practiced with computing devices similar or identical to that illustrated in
The systems and methods disclosed herein are not inherently related to any particular computer, electronic control unit, or other apparatus and may be implemented by a suitable combination of hardware, software, and/or firmware. Software implementations may include one or more computer programs comprising executable code/instructions that, when executed by a processor, may cause the processor to perform a method defined at least in part by the executable instructions. The computer program can be written in any form of programming language, including compiled or interpreted languages, and can be deployed in any form, including as a standalone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. Further, a computer program can be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a communication network. Software embodiments may be implemented as a computer program product that comprises a non-transitory storage medium configured to store computer programs and instructions, that, when executed by a processor, are configured to cause the processor to perform a method according to the instructions. In certain embodiments, the non-transitory storage medium may take any form capable of storing processor-readable instructions on a non-transitory storage medium. A non-transitory storage medium may be embodied by a compact disk, digital-video disk, hard disk drive, a magnetic tape, a magnetic disk, flash memory, integrated circuits, or any other non-transitory digital processing apparatus or memory device.
Although the foregoing has been described in some detail for purposes of clarity, it will be apparent that certain changes and modifications may be made without departing from the principles thereof. It will be appreciated that these systems and methods are novel, as are many of the components, systems, and methods employed therein. It should be noted that there are many alternative ways of implementing both the processes and apparatuses described herein. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the inventive body of work is not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.
This application is a Continuation of U.S. application Ser. No. 17/066,804, filed Oct. 9, 2020, which is a Continuation of U.S. application Ser. No. 16/272,421, filed Feb. 11, 2019, now U.S. Pat. No. 10,827,294, which is a Continuation of U.S. application Ser. No. 15/861,143, filed Jan. 3, 2018, now U.S. Pat. No. 10,244,340, which is a Continuation of U.S. application Ser. No. 15/250,870, filed Aug. 29, 2016, now U.S. Pat. No. 9,883,315, which is a Continuation of U.S. application Ser. No. 13/773,483, filed Feb. 21, 2013, now U.S. Pat. No. 9,438,996, which claims the benefit of priority of Provisional Application No. 61/601,529, filed Feb. 21, 2012, all of which are hereby incorporated by reference in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
5386478 | Plunkett | Jan 1995 | A |
5511129 | Craven et al. | Apr 1996 | A |
5727074 | Hildebrand | Mar 1998 | A |
6674864 | Kitamura | Jan 2004 | B1 |
6760451 | Craven et al. | Jul 2004 | B1 |
7483540 | Rabinowitz et al. | Jan 2009 | B2 |
7664276 | McKee Cooper | Feb 2010 | B2 |
7773755 | Terauchi et al. | Aug 2010 | B2 |
7869768 | Vishlitzky | Jan 2011 | B1 |
7899194 | Boesen | Mar 2011 | B2 |
7953456 | Romesburg et al. | May 2011 | B2 |
8175303 | Sawashi et al. | May 2012 | B2 |
8213637 | Brannmark et al. | Jul 2012 | B2 |
8867313 | Rivlin et al. | Oct 2014 | B1 |
9438996 | Maher et al. | Sep 2016 | B2 |
9883315 | Maher et al. | Jan 2018 | B2 |
10244340 | Maher et al. | Mar 2019 | B2 |
20030179891 | Rabinowitz et al. | Sep 2003 | A1 |
20050195988 | Tashev et al. | Sep 2005 | A1 |
20100042925 | DeMartin et al. | Feb 2010 | A1 |
20100142735 | Yoon et al. | Jun 2010 | A1 |
20120106747 | Crockett | May 2012 | A1 |
20130000464 | Kirsch | Jan 2013 | A1 |
20130163768 | Salter et al. | Jun 2013 | A1 |
20140003625 | Sheen | Jan 2014 | A1 |
20140169569 | Toivanen | Jun 2014 | A1 |
Number | Date | Country |
---|---|---|
1447624 | Oct 2003 | CN |
1349427 | Oct 2003 | EP |
2197220 | Jun 2010 | EP |
2003-324788 | Nov 2003 | JP |
2005-012784 | Jan 2005 | JP |
2007-259391 | Oct 2007 | JP |
10-2010-0066949 | Jun 2010 | KR |
Entry |
---|
International Search Report and International Written Opinion dated Jun. 4, 2013 for application No. PCT/2013/027184. |
Supplementary European Search Report dated Jul. 27, 2015 for EP Application No. 13752325.4. |
European Examination Report dated Jun. 22, 2016, for EPO Application No. 13752325.4. |
Fielder, L.D.; “Practical Limits for Room Equalization”; Audio Engineering Society 111th Convention Preprint; Sep. 21-24, 2001; New York, NY. |
First Chinese Office Action dated Jan. 18, 2016 for CN Application No. 201380021016.4. |
First Japanese Office Action and English translation dated Feb. 28, 2017 for Patent App. No. 2014-557890; 8 pages. |
Examination Report dated May 9, 2017 for European Patent Application No. 13752325.4; (5 pages). |
Second Chinese Office Action dated Nov. 8, 2016 for CN Application No. 201380021016.4. |
Number | Date | Country | |
---|---|---|---|
20220295210 A1 | Sep 2022 | US |
Number | Date | Country | |
---|---|---|---|
61601529 | Feb 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 17066804 | Oct 2020 | US |
Child | 17804455 | US | |
Parent | 16272421 | Feb 2019 | US |
Child | 17066804 | US | |
Parent | 15861143 | Jan 2018 | US |
Child | 16272421 | US | |
Parent | 15250870 | Aug 2016 | US |
Child | 15861143 | US | |
Parent | 13773483 | Feb 2013 | US |
Child | 15250870 | US |