SYSTEMS AND METHODS FOR PROVIDING AUGMENTED AUDIO

Abstract
A system for providing spatialized audio in a vehicle, including a vehicle orientation sensor outputting a vehicle orientation signal and being disposed on the vehicle and a controller configured to receive a user orientation signal output from a user orientation sensor being on a wearable that, during use, moves with a first user's head, wherein the controller is further configured to determine an orientation of the user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the user orientation signal, the controller being further configured to output to a first binaural device, according to the orientation of the user's head relative to the vehicle, a first spatial audio signal, such that the first binaural device produces a first spatial acoustic signal perceived by the user as originating from a first virtual source location within a cabin of the vehicle.
Description
BACKGROUND

This disclosure generally relates to systems and method for providing augmented audio in a vehicle cabin, and, particularly, to a method of augmenting the bass response of at least one binaural device disposed in a vehicle cabin.


SUMMARY

All examples and features mentioned below can be combined in any technically possible way.


A system for providing spatialized audio in a vehicle, includes: a vehicle orientation sensor outputting a vehicle orientation signal and being disposed on the vehicle; and a controller configured to receive a user orientation signal output from a user orientation sensor being disposed on a wearable that, during use, moves with a first user's head, wherein the controller is further configured to determine an orientation of the user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the user orientation signal, the controller being further configured to output to a first binaural device, according to the orientation of the user's head relative to the vehicle, a first spatial audio signal, such that the first binaural device produces a first spatial acoustic signal perceived by the user as originating from a first virtual source location within a cabin of the vehicle.


In an example, the system further includes an error sensor configured to detect the orientation of the user's head relative to the vehicle and to output an error sensor signal, wherein the controller is further configured to correct a drift between the user orientation signal and the vehicle orientation signal according to the orientation of the user's head detected by the error sensor.


In an example, the error sensor includes at least one of: a time-of-flight sensor, a LIDAR device, a camera, or a microphone disposed on the user's head.


In an example, the controller samples the error sensor signal at a rate slower than the rates at which the vehicle orientation signal and the user orientation signal are sampled.


In an example, the vehicle orientation sensor comprises a first plurality of sensors, wherein the user orientation sensor comprises a second plurality of sensors, wherein the controller is further configured to correct a drift between the user orientation signal and the vehicle orientation signal according to a measure of similarity between at least one sensor of the first plurality of sensors and one sensor of the second plurality of sensors.


In an example, the controller comprises a first controller and a second controller, the first controller receiving the vehicle orientation signal and the user orientation signal and determining the orientation of the user's head relative to the vehicle and outputting a position signal to the second controller, the second controller, receiving the position signal, outputting to the first spatial acoustic signal to the first binaural device.


In an example, the wearable is the first binaural device.


In an example, the system further includes a plurality of speakers disposed in a perimeter of a cabin of the vehicle, wherein the first spatial audio signal comprises at least an upper range of a first content signal, wherein the controller is further configured to drive the plurality of speakers with a driving signal such that a first bass content of the first content signal is produced in the cabin.


In an example, the controller is configured to receive a second user orientation signal output from a second user orientation sensor being disposed on a second wearable that, during use, moves with a second user's head, wherein the controller is further configured to determine an orientation of the second user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the second user orientation signal, the controller being further configured to output to a second binaural device, according to the orientation of the second user's head relative to the vehicle, a second spatial audio signal, such that the second binaural device produces a second spatial acoustic signal perceived by the second user as originating from the first or a second virtual source location within a cabin of the vehicle.


In an example, the second spatial audio signal comprises at least an upper range of a second content signal, wherein the controller is further configured to drive the plurality of speakers in accordance with a first array configuration such that the first bass content is produced in a first listening zone within the cabin and in accordance with a second array configuration such that a second bass content of the second content signal produced in a second listening zone within the cabin, wherein in the first listening zone a magnitude of the first bass content is greater than a magnitude of the second bass content and in the second listening zone the magnitude of the second bass content is greater than the magnitude of the first bass content.


In an example, the vehicle orientation sensor is: integrated with the vehicle or brought into and fixed to the vehicle by a user.


A method for providing spatialized audio in a vehicle, includes: receiving a user orientation signal output from a user orientation sensor being disposed on a wearable that, during use, moves with a first user's head; receiving a vehicle orientation signal from a vehicle orientation sensor being disposed on the vehicle; determining an orientation of the user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the user orientation signal; and outputting to a first binaural device, according to the orientation of the user's head relative to the vehicle, a first spatial audio signal, such that the first binaural device produces a first spatial acoustic signal perceived by the user as originating from a first virtual source location within a cabin of the vehicle.


In an example, the method further includes receiving an error sensor signal output from an error sensor configured to detect the orientation of the user's head relative to the vehicle; and correcting a drift between the user orientation signal and the vehicle orientation signal according to the orientation of the user's head detected by the error sensor.


In an example, the error sensor includes at least one of: a time-of-flight sensor, a LIDAR device, a camera, a microphone disposed on the user's head.


In an example, the error sensor signal is sampled at a rate slower than the rates at which the vehicle orientation signal and the user orientation signal are sampled.


In an example, the method further includes the step of correcting a drift between the user orientation signal and the vehicle orientation signal, wherein the vehicle orientation sensor comprises a first plurality of sensors, wherein the user orientation sensor comprises a second plurality of sensors, wherein the drift is corrected according to a measure of similarity between at least one sensor of the first plurality of sensors and one sensor of the second plurality of sensors.


In an example, the wearable is the first binaural device.


In an example, the method further includes driving a plurality of speakers with a driving signal such that a first bass content of the spatial audio signal is produced in the cabin.


In an example, the method further includes receiving a second user orientation signal output from a second user orientation sensor being disposed on a second wearable that, during use, moves with a second user's head, determining an orientation of the second user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the second user orientation signal, outputting to a second binaural device, according to the orientation of the second user's head relative to the vehicle, a second spatial audio signal, such that the second binaural device produces a second spatial acoustic signal perceived by the second user as originating from the first or a second virtual source location within a cabin of the vehicle.


In an example, the method further includes driving the plurality of speakers in accordance with a first array configuration such that the first bass content is produced in a first listening zone within the cabin and in accordance with a second array configuration such that a second bass content of the second spatial audio signal is produced in a second listening zone within the cabin, wherein in the first listening zone a magnitude of the first bass content is greater than a magnitude of the second bass content and in the second listening zone the magnitude


The details of one or more implementations are set forth in the accompanying drawings and the description below. Other features, objects, and advantages will be apparent from the description and the drawings, and from the claims.





BRIEF DESCRIPTION OF THE DRAWINGS

In the drawings, like reference characters generally refer to the same parts throughout the different views. Also, the drawings are not necessarily to scale, emphasis instead generally being placed upon illustrating the principles of the various aspects.



FIG. 1A depicts an audio system for providing augmented audio in a vehicle cabin, according to an example.



FIG. 1B depicts an audio system for providing augmented audio in a vehicle cabin, according to an example.



FIG. 2 depicts an open-ear wearable, according to an example.



FIG. 3 depicts an open-ear wearable, according to an example.



FIG. 4 depicts a flowchart of a method for providing augmented audio in a vehicle cabin, according to an example.



FIG. 5 depicts an audio system for providing augmented spatialized audio in a vehicle cabin, according to an example.



FIG. 6 depicts a flowchart of a method for providing augmented spatialized audio in a vehicle cabin, according to an example.



FIG. 7A depicts a cross-over plot according to an example.



FIG. 7B depicts a cross-over plot according to an example.



FIG. 8A depicts an audio system for providing augmented spatialized audio in a vehicle cabin, according to an example.



FIG. 8B depicts an audio system for providing augmented spatialized audio in a vehicle cabin, according to an example.



FIG. 8C depicts an audio system for providing augmented spatialized audio in a vehicle cabin, according to an example.



FIG. 8D depicts an audio system for providing augmented spatialized audio in a vehicle cabin, according to an example.



FIG. 9A depicts a representation of a user orientation and vehicle orientation as compared to a measured user orientation and vehicle orientation.



FIG. 9B depicts a representation of a user orientation and vehicle orientation as compared to a measured user orientation and vehicle orientation.



FIG. 9C depicts a representation of a user orientation and vehicle orientation as compared to a measured user orientation and vehicle orientation.



FIG. 10A depicts a representation of a wearable having microphones disposed on opposing sides, such as a delay between receipt of an audio signal can be detected.



FIG. 10B depicts a representation of a wearable having microphones disposed on opposing sides, such as a delay between receipt of an audio signal can be detected.



FIG. 11A depicts of representation of accelerations detected by separate accelerometers of a user orientation sensor.



FIG. 11B depicts of representation of accelerations detected by separate accelerometers of a user orientation sensor.



FIG. 12 depicts a flowchart for a method of providing spatialized audio to a binaural device within a vehicle according to the orientation of a user's head relative to a vehicle.





DETAILED DESCRIPTION

A vehicle audio system that includes only perimeter speakers is limited in its ability to provide different audio content to different passengers. While the vehicle audio system can be arranged to provide separate zones of bass content with satisfactory isolation, this cannot be similarly said about upper range content, in which the wavelengths are too short to adequately create separate listening zones with independent content using the perimeter speakers alone.


The leakage of upper-range content between listening zones can be solved by providing each user with a wearable device, such as headphones. If each user is wearing a pair of headphones, a separate audio signal can be provided to each user with minimal sound leakage. But minimal leakage comes at the cost of isolating each passenger from the environment, which is not desirable in a vehicle context. This is particularly true of the driver, who needs to be able to hear sounds in the environment such as those produced by emergency vehicles or the voices of the passengers, but it is also true of the rest of the passengers which typically want to be able to engage in conversation and interact with each other.


This can be resolved by providing each user with a binaural device such as an open-ear wearable or near-field speakers, such as headrest speakers, that provides each passenger with separate upper range audio content while maintaining an open path to the user's ears, allowing users to engage with their environment. But open-ear wearables and near-field speakers typically do not provide adequate bass response in a moving vehicle as the road noise tends to mask the same frequency band.


Turning now to FIG. 1A there is shown a schematic view representative of the audio system for providing augmented audio in a vehicle cabin 100. As shown, the vehicle cabin 100 includes a set of perimeter speakers 102. (For the purposes of this disclosure a speaker is any device receiving an electrical signal and transducing it into an acoustic signal.) A controller 104, disposed in the vehicle, is configured to receive a first content signal u1 and a second content signal u2. The first content signal u1 and second content signal u2 are audio signals (and can be received as analog or digital signals according to any suitable protocol) that each include a bass content (i.e., content below 250 Hz±150 Hz) and an upper range content (i.e., content above 250 Hz±150 Hz). The controller 104 is configured to drive perimeter speakers 102 with driving signals d1— d4 to form at least a first array configuration and a second array configuration. The first array configuration, formed by at least a subset of perimeter speakers 102, constructively combines the acoustic energy generated by perimeter speakers 102 to produce the bass content of the first content signal u1 in a first listening zone 106 arranged at a first seating position P1. The second array configuration, similarly formed by at least a subset of perimeter speakers 102, constructively combines the acoustic energy generated by perimeter speakers 102 to produce the bass content of the second content signal u2 in a second listening zone 108 arranged at a second seating position P2. Furthermore, the first array configuration can destructively combine the acoustic energy generated by perimeter speakers 102 to form a substantial null at the second listening zone 108 (and any other seating position within the vehicle cabin) and the second array configuration can destructively combine the acoustic energy generated by perimeter speakers 102 to form a substantial null at the first listening zone (and any other seating position within the vehicle cabin).


It should be understood that in various examples there can be some or total overlap between the subsets of perimeter speakers 102 arrayed to produce the bass content of the first content signal u1 in the first listening zone 106 and the subsets of perimeter speakers 102 arrayed to produce the bass content of the second content signal u2 in the second listening zone.


Given a substantially same magnitude of bass content in the first and second content signals, arraying of the perimeter speakers 102 means that the magnitude of the bass content of the first content signal u1 is greater in the first listening zone 106 than the magnitude of the bass content of the second content signal u2. Similarly, the magnitude of the bass content of the second content signal u2 is greater than the magnitude of the bass content of the first content signal u1. The net effect is that a user seated at position P1 primarily perceives the bass content of the first content signal u1 as greater than the bass content of the second content signal u2, which may not be perceived at in some instances. Similarly, a user seated at position P2 primarily perceives the bass content of the second content signal u2 as greater than the bass content of the first content signal u1. In one example, the magnitude of the bass content of the first content signal u1 is greater than the magnitude of the bass content of the second content signal u2 by at least 3 dB in the first listening zone, and, likewise, the magnitude of the bass content of the second content signal u2 is greater than the magnitude of the bass content of the first content signal u1 by at least 3 dB in the second listening zone.


Although only four perimeter speakers 102 are shown, it should be understood that any number of perimeter speakers 102 greater than one can be used. Furthermore, for the purposes of this disclosure the perimeter speakers 102 can be disposed in or on the vehicle doors, pillars, ceiling, floor, dashboard, rear deck, trunk, under seats, integrated within seats, or center console in the cabin 100, or any other drive point in the structure of the cabin that creates acoustic bass energy in the cabin.


In various examples, the first content signal u1 and second content signal u2 (and any other received content signals) can be received from one or more of a mobile device (e.g., via a Bluetooth connection), a radio signal, a satellite radio signal, or a cellular signal, although other sources are contemplated. Furthermore, each content signal need not be received contemporaneously but rather can have been previously received and stored in memory for playback at a later time. Furthermore, as mentioned above, the first content signal u1 and second content signal u2 can be received as an analog or digital signal according to any suitable communications protocol. In addition, because the first content signal u1 and second content signal u2 can be transmitted digitally, which is comprised of a set of binary values, the bass content and upper range content of these signals refers to the constituent signals of the respective frequency ranges of the bass content and upper range content when the content signal is converted into an analog signal before being transduced by a speaker or other device.


As shown in FIG. 1A, binaural devices 110 and 112 are respectively positioned to produce a stereo first acoustic signal 114 in the first listening zone 106 and a stereo second acoustic signal 116 in the second listening zone. As shown in FIG. 1A, binaural device 110 and 112 are comprised of speakers 118, 120 disposed in a respective headrest disposed proximate to listening zones 106, 108. Binaural device 110, for example, comprises left speaker 118L, disposed in a headrest to deliver left-side first acoustic signal 114L to the left ear of a user seated in the first seating position P1 and a right speaker 118R to deliver right-side first acoustic signal 114R to the right ear of the user. In the same way, binaural device 112 comprises left speaker 120L disposed in a headrest to deliver left-side second acoustic signal 116L to the left ear of a user seated in the second seating position P2 and right speaker 120R to deliver right-side second acoustic signal 116R to the right ear of the user. Although the acoustic signals 114, 116 are shown as comprising left and right stereo components, it should be understood that in some examples, one or both acoustic signals 114, 116 could be mono signals, in which both the left side and right side are the same. Binaural device 110, 112 can each further employ a set of cross-cancellation filters that cancel the audio on each respective side produced by opposite side. Thus, for example, binaural device 110 can employ a set of cross-cancellation filters to cancel at the user's left ear audio produced for the user's right ear and vice versa. In examples in which the binaural device is a wearable (e.g., an open-ear headphone) and has drive points close to the ears, crosstalk cancellation is typically not required. However, in the case of headrest speakers or wearables that are further away (e.g., Bose SoundWear), the binaural device would typically employ some measure crosstalk cancellation to achieve binaural control.


Although the first binaural device 110 and second binaural device 112 are shown as speakers disposed in a headrest, it should be understood that the binaural devices described in this disclosure can be any device suitable for delivering to the user seated at the respective position, independent left and right ear acoustic signals (i.e., a stereo signal). Thus, in an alternative example, the first binaural device 110 and/or second binaural device 112 could be comprised of speakers located in other areas of vehicle cabin 100 such as the upper seatback, headliner, or any other place that is disposed near to the user's ears, suitable for delivering independent left and right ear acoustic signals to the user. In yet another alternative example, first binaural device 110 and/or second binaural device 112 can be an open-ear wearable worn by the user seated at the respective seating position. For the purposes of this disclosure, an open-ear wearable is any device designed to be worn by a user and being capable of delivering independent left and right ear acoustic signals while maintaining an open path to the user's ear. FIGS. 2 and 3 show two examples of such open ear wearables. The first open ear wearable is a pair of frames 200, featuring a left speaker 202L and a right speaker 202R located in the left temple 204L and right temple 204R, respectively. The second is a pair of open-ear headphones 300 featuring a left speaker 302L and a right speaker 302R. Both frames 200 and open-ear headphones 300 retain an open path to the user's ear, while being able to provide separate acoustic signals to the user's left and right ears.


Controller 104 can provide at least the upper range content of the first content signal u1 via binaural signal b1 to the first binaural device 110 and at least the upper range content of the second signal content signal u2 via binaural signal b2 to the second binaural device 112. (In an example, the entire range, including the bass content, of the first content signal u1 and second content signal u2 is respectively delivered to the first binaural device 110 and second binaural device 112.) As a result, the first acoustic signal 114 comprises at least the upper range content of the first content signal u1 and the second acoustic signal 116 comprises at least the upper range content of the second signal u2. The production of the bass content of the first content signal u1 in the first listening zone 106 by perimeter speaker 102 augments the production of the upper range content of the first signal u1 produced by the first binaural device 110, and the production of the bass content of the second content signal u2 in the second listening zone 108 by perimeter speakers 102 augments the production of the upper range content of the second content signal u2 produced by the second binaural device.


A user seated at seating position P1 thus perceives the first content signal u1 played in the first listening zone 106 from the combined outputs of the first arrayed configuration of perimeter speakers 102 and first binaural device 110. Likewise, the user seated at seating position P2 perceives the second content signal u2 played in the second listening zone 108 from the combined outputs of the second arrayed configuration of perimeter speakers 102 and second binaural device 112.



FIGS. 7A and 7B depict example plots of frequency cross-over between bass content and upper range content of an example content signal (e.g., first content signal u1) at 100 Hz and 200 Hz respectively. As described above, the cross-over between the bass content and upper range content can occur at, e.g., 250 Hz±150 Hz, thus the crossover 100 Hz or 200 Hz are examples of this range. As shown, the combined total response at the listening zone is perceived to be a flat response. (Of course, the flat response is only one example of a frequency response, and other examples can, e.g., boost the bass, midrange, and/or treble, depending on the desired equalization.)


Binaural signals b1, b2 (and any other binaural signals generated for additional binaural devices) are generally N-channel signals, where N≥2 (as there is at least one channel per ear). N can correlate to the number of speakers in the rendering system (e.g., if a headrest has four speakers, the associated binaural signal typically has four channels). In instances in which the binaural device employs crosstalk cancellation, there may exist some overlap between content in the channels in the for the purposes of cancellation. Typically, though, the mixing of signals is performed by a crosstalk cancellation filter disposed within the binaural device, rather than in the binaural signal received by the binaural device.


Controller 104 can provide binaural signals b1, b2 in either a wired or wireless manner. For example, where binaural device 110 or 112 is an open-ear wearable, the respective binaural signal b1, b2 can be transmitted over Bluetooth, WiFi, or any other suitable wireless protocol.


In addition, controller 104 can be further configured to time-align the production of the bass content in the first listening zone 106 with the production of the upper range content by the first binaural device 110 to account for the wireless, acoustical, or other transmission delays intrinsic to the production of such signals. Similarly, the controller 104 can be further configurated to time-align the production of the bass content in the second listening zone 108 with the production of the upper range content by the second binaural device 112. There will be some intrinsic delay between the output of driving signals d1-d4 and the point in time that the bass content, transduced by perimeter speakers 102, arrives at the respective listening zone 106, 108. The delay comprises the time required for driving signal d1-d4 to be transduced by the respective speaker 102 into an acoustic signal, and to travel to the first listening zone 106 or the second listening 108 from the respective speaker 102. (Although it is conceivable that other factors could influence the delays.) Because each perimeter speaker 102 is likely located some unique distance from the first listening zone 106 and the second listening zone 108, the delay can be calculated for each perimeter speaker 102 separately. Furthermore, there will be some delay between outputting binaural signals b1, b2 and the respective production of acoustic signals 114, 116 in the first listening zone 106 and second listening zone 108. This delay will be a function of the time to process the received binaural signal b1, b2 (in the event that the binaural signal is encoded in a communication protocol, such as a wireless protocol, and/or where binaural device performs some additional signal processing) and to transduce the binaural signal b1, b2 into acoustic signals 114, 116, and the time for the acoustic signals 114, 116 to travel to the user seated at position P1, P2 (although, because each binaural device is located relatively near to the user, this is likely negligible). (Again, other factors could influence the delay.) Thus, taking these delays into account, controller 104 can time the production of driving signals d1-d4 and binaural signals b1, b2 such that the production, by perimeter speakers 102, of the bass content of first content signal u1 is time-aligned in the first listening zone 106 with the production, by the first binaural device 110, of the upper range content of the first content signal u1, and the production, by perimeter speakers 102 of the bass content of the second content signal u2 is time-aligned in the second listening zone 108 with the production, by the second binaural device 112, of the upper range of the second content signal u2.


For the purposes of this disclosure, “time-aligned” refers to the alignment in time of the production of the bass content and upper range content of a given content signal at given point in space (e.g., a listening zone), such that, at the given point in space, the content is accurately reproduced. It should be understood that the bass content and upper range content need only be time aligned to a degree sufficient for a user to perceive the content signal is accurately reproduced. Generally, an offset of 90° at the crossover frequency between the bass content and upper range content is acceptable in a time-aligned acoustic signal. To provide a couple of examples at several different crossover frequencies, an acceptable offset could be +/−2.5 ms for 100 Hz, +/−1.25 ms for 200 Hz, +/−1 ms for 250 Hz, and +/−0.625 ms for 400 Hz. However, it should be understood that, for the purposes of this disclosure, anything up to a 180° offset at the crossover frequency is considered time aligned.


As shown in FIGS. 7A and 7B, there is additional overlap between the bass content and upper range content beyond the cross-over frequency. The phase of these frequencies within the overlap can be individually shifted to align the upper range content and bass content in time; as will be understood, the phase shift applied will be dependent on frequency. For example, one or more all-pass filters can be included, designed to introduce a phase shift, at least to the overlapping frequencies of the upper range content and the bass content, in order to achieve the desired time-alignment across frequency.


The time alignment can be a priori established for a given binaural device. In the example of headrest speakers, the delay between receiving the binaural signal and producing the acoustic signal will always be the same and the delays can thus be set as a factory setting. However, where the binaural device 110, 112 is a wearable, the delay will typically vary from wearable to wearable, based on the varied times required to process the respective binaural signal b1, b2, and to produce the acoustic signal 114, 116 (this is especially true in the case of wireless protocols which have notoriously variable latency). Accordingly, in one example, controller 104 can store a plurality of delay presets for time-aligning the production of the bass content with the production of the acoustic signal 114, 116 for various wearable devices or types of wearable devices. Thus, when controller 104 connects to a particular wearable device it can identify the wearable (e.g., a pair of Bose Frames) and retrieve from storage a particular prestored delay for time-aligning the bass content with acoustic signal 114, 116 produced by the identified wearable. In an alternative example, a prestored delay can be associated with a particular device type. For example, if the delays associated with wearables operating a particular communication protocol (e.g., Bluetooth) or protocol version (e.g., a Bluetooth version) are typically the same, controller 104 can select delay according to the detected communication protocol or communication protocol version. These prestored delays for a given device or type of device can be determined by employing a microphone at a given listening zone and calibrating the delay, manually or by an automated process, until the bass content of a given content signal is time-aligned with the acoustic signal of a given binaural device at the listening zone. In yet another example, the delays can be calibrated according to a user input. For example, a user wearing the open-ear wearable can sit in a seating position P1 or P2 and adjust the production of drive signal d1-d4 and/or binaural signals b1, b2 until the bass content is correctly time-aligned with the upper range of acoustic signal 114, 116. In another example, the device can report to controller 104 a delay necessary for time-alignment.


In alternative examples, the time alignment can be determined automatically during runtime, rather than by a set of prestored delays. In an example, a microphone can be disposed on or near the binaural device (e.g., on a headrest or on the wearable) and used to produce a signal to the controller to determine the delay for time alignment. One method for automatically determining time-alignment is described in US 2020/0252678, titled “Latency Negotiation in a Heterogeneous Network of Synchronized Speakers” the entirety of which is herein incorporated by reference, although any other suitable method for determining delay can be used.


As described above, the time alignment can be achieved across a range of frequencies using an all-pass filter(s). To account for the different delays of various binaural devices, the particular filter(s) implemented can be selected from a set of stored filters, or the phase change implemented by the all-pass filter(s) can be adjusted. The selected filter or the phase change can, as described above, be based upon different devices or device types, by a user input, according to a delay detected by microphones on the wearable device, according to a delay reported by the wearable device, etc.


In the example of FIG. 1A, controller 104 generates both driving signals d1-d4 and binaural signal b1, b2. In alternative example, however, one or more mobile devices can provide the binaural signals b1, b2. For example, as shown in FIG. 1B, a mobile device 122 provides binaural signal b1 to binaural device 110 (e.g., where the binaural device 110 is an open-ear wearable) via a wired or wireless (e.g., Bluetooth) connection. For example, a user can enter the vehicle cabin 100 wearing the open-ear wearable binaural device 110 and listening to music via a paired Bluetooth connection (binaural signal b 1) with mobile device 122. Upon entering vehicle cabin 100, controller 104 can begin to provide the bass content of first content signal u1 while mobile device 122 continues to provide binaural signal b1 to the open ear wearable binaural device 110. In this example, controller 104 can receive from the mobile device 122 first content signal u1 in order to produce the bass content of first content signal u1 in the first listening zone 106. Thus, mobile device 122 can pair with (or otherwise be connected to) both binaural device 110 and controller 104 to provide binaural signal b1 and first content signal u1. In an alternative example, mobile device 122 can broadcast a single signal that is received by both controller 104 and binaural device 110 (in this example, each device can apply a respective high-pass/low-pass for crossover). For example, the Bluetooth 5.0 standard provides such an isochronous channel for locally broadcasting a signal to nearby devices. In an alternative example, rather than transmitting first content signal u1, mobile device 122 can transmit to controller 104 metadata of the content transmitted to the first binaural device 110 by first binaural signal b1, allowing controller 104 to source the correct first content signal u1 (i.e., the same content) from an outside source such as a streaming service.


While only one mobile device 122 is shown in FIG. 1B, it should be understood that any number of mobile devices can provide binaural signals to any number of binaural devices (e.g., binaural devices 110, 112) disposed in the vehicle cabin 100.


Of course, as described in connection with FIG. 1B, controller 104 can receive first content signal u1 from a mobile device. Thus, in one example, a user can be wearing open-ear wearable first binaural device 110 when entering the vehicle, at which time, the mobile device 122 ceases transmitting content to the first binaural device and instead provides first content signal u1 to controller 104 which assumes transmitting binaural signal b1, e.g., through a wireless connection such as Bluetooth. Similarly, for multiple binaural devices (e.g., binaural devices 110, 112), receiving signals from multiple mobile devices, controller 104 can assume transmitting a respective binaural signal (e.g., binaural signals b1, b2) to the binaural device, rather than the mobile device.


Controller 104 can comprise a processor 124 (e.g., a digital signal processor) and a non-transitory storage medium 126 storing program code that, when executed by processor 124, carries out the various functions and methods described in this disclosure. It should, however, be understood that, in some examples, controller 104, can be implemented as hardware only (e.g., as an application-specific integrated circuit or field-programmable gate array) or as some combination of hardware, firmware, and software.


In order to array perimeter speakers 102 to provide bass content to first listening zone 106 and second listening zone 108, controller 104 can implement a plurality of filters that each adjust the acoustic output of perimeter speakers 102 so that the bass content of the first content signal u1 constructively combines at the first listening zone 106 and the bass content of the second signal u2 constructively combines at the second listening zone 108. While such filters are normally implemented as digital filters, these filters could alternatively be implemented as analog filters.


In addition, although only two listening zones 106 and 108 are shown in FIGS. 1A and 1B, it should be understood that controller 104 can receive any number of content signals and create any number of listening zones (including only one) by filtering the content signals to array perimeter speakers, each listening zone receiving the bass content of a unique content signal. For example, in a five-seat car, the perimeter speakers can be arrayed to produce five separate listening zones, each producing the bass content of a unique content signal (i.e., in which the magnitude of the bass content for the respective content signal is loudest, assuming that the bass contents of each content signal are played at substantially equal magnitude in other listening zone). Furthermore, a separate binaural device can be disposed at each listening zone and receive a separate binaural signal, augmented by and time-aligned with the bass content produced in the respective listening zone.


In the above examples, binaural devices 110, 112 (or any other binaural devices) can deliver to both users the same content. In this example, controller 104 can augment the acoustic signal produced by the binaural devices with bass content produced by perimeter speakers 102 without creating separate listening zones for playing separate content. The bass content can be time-aligned with the upper range content played from both binaural devices 110, 112, thus both users perceive the played content signal, including the upper range signal delivered by the binaural devices 110, 112 and the bass content played by perimeter speakers 102. Although each device receives the same program content signal, it is conceivable that the user would select different volume levels of the same content. In this case, rather than creating separate listening zones, controller 104 can employ the first array configuration and second array configuration to create separate volume zones, in which each user perceives the same program content at different volumes.


In an example, it is not necessary that each user have the same have an associated binaural device, rather some users can listen only to the content produced by the perimeter speakers 102. For this example, the perimeter speakers 102 would produce not only the bass content, but also the upper range content of the program content signal (e.g., program content signal u1). For the user's with binaural devices, the program content signal is perceived as a stereo signal, as provided for by the binaural signal (e.g., binaural signal b1) and by virtue of the left and right speakers of the binaural device. Indeed, it should be understood that, in each of the examples described in this disclosure, there may be some or complete overlap in spectral range between the signals produced by the perimeter speakers 102 and the binaural devices (e.g., binaural devices 110, 112). Those with binaural devices having an overlap in spectral range with the perimeter speakers 102 receive an enhanced experience with improved stereo, audio staging, and perceived spaciousness.


It should be understood that navigation prompts and phone calls are among the program content signals that can be directed toward particular users in listening zones. Thus, a driver can hear navigation prompts produced by a binaural device (e.g., binaural device 110) with bass augmented by the perimeter speakers while the passengers listen to music in a different listening zone.


In addition, the microphones on wearable binaural devices can be used for voice pick-up, for traditional uses such as phone call, vehicle-based or mobile device-based voice recognition, digital assistants, etc.


Further, rather than one set of filters, a plurality of filters can be implemented by controller 104 depending on the configuration of the vehicle cabin 100. For example, various parameters within the cabin will change the acoustics of the vehicle cabin 100, including, the number of passengers in the vehicle, whether the windows are rolled up or down, the position of the seats in the vehicle (e.g., whether the seats are upright or reclined or moved forward or back in the vehicle cabin), etc. These parameters can be detected by controller 104 (e.g., by receiving a signal from the vehicles on-board computer) and implement the correct set of filters to provide the first, second, and any additional arrayed configurations. Various sets of filters, for example, can be stored in memory 126 and retrieved according to the detected cabin configuration.


In an alternative example, the filters can be a set of adaptive filters that are adjusted according to a signal received from an error microphone (e.g., disposed on binaural device or otherwise within a respective listening zone) in order to adjust the filter coefficients to align the first listening zone over a respective seating position (first seating position P1 or second seating position P2), or to adjust for changing cabin configurations, such as whether the windows are rolled up or down.



FIG. 4 depicts a flowchart for a method 400 of providing augmented audio to users in a vehicle cabin. The steps of method 400 can be carried out by a controller (such as controller 104) in communication with a set of perimeter speakers (such as perimeter speakers 102) disposed in a vehicle and further in communication with a set of binaural devices (such as binaural device 110, 112) disposed at respective seating positions within the vehicle.


At step 402 a first content signal and second content signal are received. These content signals can be received from multiple potential sources such as mobile devices, radio, satellite radio, a cellular connection, etc. The content signals each represent audio that may include a bass content and an upper range content.


At steps 404 and 406 a plurality of perimeter speakers are driven in accordance with a first array configuration (step 404) and a second array configuration (step 406) such that the bass content of the first content signal is produced in a first listening zone and the bass content of the second content signal is produced in a second listening zone in the cabin. The nature of the arraying produces listening zones such that, when the bass content of the first content signal is played in the first listening zone at the same magnitude as the bass content of the second signal is played in the second listening zone, the magnitude of the bass content of the first content signal will be greater than the magnitude of the bass content of the second content signal (e.g., by at least 3 dB) in the first listening zone, and the magnitude of the bass content of the second signal will be greater than the magnitude of the bass content of the first content signal (e.g., by at least 3 dB) in the second listening zone. In this way, a user seated at the first seating position will perceive the magnitude of the first bass content as greater than the second bass content. Likewise, a user seated at the second seating position will perceive the magnitude of the second bass content as greater than the first bass content.


At steps 408 and 410 the upper range content of the first content signal is provided to a first binaural device positioned to produce the upper range content in the first listening zone (step 408) and the upper range content of the second content signal is provided to a second binaural device positioned to produce the upper range content in the second listening zone (step 410). The net result is a user seated at the first seating position perceives the first content signal from the combination of outputs of the first binaural device and the perimeter speakers and a user seated at the second seating position perceives the second content signal from the combination of outputs of the second binaural device and the perimeter speakers. Stated differently, the perimeter speakers augment the upper range of the first content signal as produced by the first binaural device with the bass of the first content signal in the first listening zone, and augment the upper range of the second content signal as produced by the second binaural signal with the bass of the second content signal in the second listening zone. In various alternative examples, the first binaural device is an open-ear wearable or speakers disposed in a headrest.


Furthermore, the production of the bass content of the first content signal in the first listening zone can be time-aligned with the production of the upper range of the first content signal by the first binaural device in the first listening zone and the production of the second bass content in the second listening zone can be time-aligned with the production of the upper range of the second content signal by the second binaural device. In an alternative example, the first upper range content or second upper range content can be provided to the first binaural device or second binaural device by a mobile device, with which the production of the bass content is time-aligned.


Although method 400 is described for two separate listening zones and two binaural devices, it should be understood that method 400 can be extended to any number of listening zones (including only one) disposed within the vehicle and at which a respective binaural device is disposed. In the case of a single binaural device and listening zone, isolation to other seats is no longer important and the plurality of perimeter speaker filters can be different from the multi-zone case in order to optimize for bass presentation. (The case of a single user can, for example, be determined by a user interface or through sensors disposed in the seats.)


Turning now to FIG. 5 there is shown an alternative schematic of a vehicle audio system disposed in a vehicle cabin 100, in which perimeter speakers 102 are employed to augment the bass content of at least one binaural device producing spatialized audio. In this example, controller 504 (an alternative example of controller 104) is configured to produce binaural signals b1, b2 as spatial audio signals that cause binaural device 110 and 112 to produce acoustic signals 114, 116 as spatial acoustic signals, perceived by a user as originating from a virtual audio source, SP1 and SP2 respectively. Binaural signal b1 is produced as spatial audio signals according to the position of the head of a user seated at position P1. Similarly, binaural signal b2 is produced as spatial audio signals according to the position of the head of a user seated at position P2. Similar to the example of FIGS. 1A and 1B, these spatialized acoustic signals, produced by binaural devices 110, 112, can be augmented by bass content produced by the perimeter speakers 102 and driven by controller 504.


As shown in FIG. 5, a first headtracking device 506 and a second headtracking device 508 are disposed to respectively detect the position of the head of a user seated at seating position P1 and a user seated at seating position P2. In various examples, the first headtracking device 506 and second headtracking device 508 can be comprised of a time-of-flight sensor configured to detect the position of a user's head within the vehicle cabin 100. However, a time-of-flight sensor is only possible example. Alternatively, multiple 2D cameras that triangulate on the distance from one of the camera focal points using epi-polar geometry, such as the eight-point algorithm, can be used. Alternatively, each headtracking device can comprise a LIDAR device, which produces a black and white image with ranging data for each pixel as one data set. In alternative examples, where each user is wearing an open-ear wearable, the headtracking can be accomplished, or may be augmented, by tracking the respective position of the open-ear wearable on the user, as this will typically correlate to the position of the user's head. In still other alternative examples, capacitive sensing, inductive sensing, inertial measurement unit tracking in combination with imaging, can be used. It should be understood that the above-mentioned implementations of headtracking device are meant to convey that a range of possible devices and combinations of devices might be used to track the location of a user's head.


For the purposes of this disclosure, detecting the position of a user's head can comprise detecting any part of the user, or of a wearable worn by the user, from which the position of the center of user's cranium can be derived. For example, the location of the user's ears can be detected, from which a line can be drawn between the tragi to find the middle in approximation of the finding the center. Detecting the position of the user's head can also including detecting the orientation of the user's head, which can be derived according to any method for finding the pitch, yaw, and roll angles. Of these, the yaw is particularly important as it typically affects the ear distance to each binaural speaker the most.


First headtracking device 506 and second headtracking device 508 can be in communication with a headtracking controller 510 which receives the respective outputs h1, h 2 of first headtracking device 506 and second headtracking device 508 and determines from them the position of the user's head seated at position P1 or position P2 and generates an output signal to controller 504 accordingly. For example, headtracking controller 510 can receive raw output data h1 from first headtracking device 506, interpret the position of the head of a user seated at position P1 and output a position signal e1 to controller 504 representing the detected position. Likewise, headtracking controller 510 can receive output data h2 from second headtracking device 508 and interpret the position of the head of a user seated at seating position P2 and output a position signal e2 to controller 504 representing the detected position. Position signals e1 and e2 can be delivered real-time as coordinates that represent the position of the user's head (e.g., including the orientation as determined by pitch, yaw, and roll).


Controller 510 can comprise a processor 512 and non-transitory storage medium 514 storing program code that, when executed by processor 512 performs the various functions and methods disclosed herein for producing the position signal, including receiving the output signal of each headtracking device 506, 508 and for generating the position signal e1, e 2 to controller 104. In an example, controller 510 can determine the position of user's head through stored software or with a neural network that has been trained to detect the position of the user's head according to the output of a headtracking device. In an alternative example, each headtracking device 506, 130, can comprise its own controller for carrying out the functions of controller 510. In yet another example, controller 504 can receive the outputs of headtracking devices 506, 508 directly and perform the processing of controller 510.


Controller 504, receiving the position signal e1 and/or e2 can generate binaural signal b1 and/or b2 such that at least one of binaural device 110, 112 generates an acoustic signal that is perceived by a user as originating at some virtual point in space within the vehicle cabin 100 other than the actual location of the speakers (e.g., speakers 118, 120) generating the acoustic signal. For example, controller 504 can generate a binaural signal b1 such that binaural device 110 generates an acoustic signal 114 perceived by a user seated at seating position P1 as originating at spatial point SP1 (represented in FIG. 5 in dotted lines as this is a virtual sound source). Similarly, controller 504 can generate a binaural signal b2 such that binaural device 112 generates an acoustic signal 116 perceived by a user seated at seating position P2 as originating at spatial point SP2. This can be accomplished by filtering and/or attenuating the binaural signals b1, b2 according to a plurality of head-related transfer functions (HRTFs) which adjust acoustic signals 114, 116 to slate sound from the virtual spatial point (e.g., spatial point SP1, SP2). As the signals are binaural, i.e., relate to both of the listener's ears, the system can utilize one or more HRTFs to simulate sound specific to various locations around the listener. It should be appreciated that the particular left and right HRTFs used by the controller 504 can be chosen based on a given combination of azimuth angle and elevation detected between the relative position of the user's left and right ears and the respective spatial position SP1, SP2. More specifically, a plurality of HRTFs can be stored in memory and be retrieved and implemented according to the detected position of the user's left and right ears and selected spatial position SP1, SP2. However, it should be understood that, where binaural device 110, 112 is an open-ear wearable, the location of the user's ears can be substituted for or determined from the location of the open-ear wearable.


Although two different spatial points SP1, SP2 are shown in FIG. 5, it should be understood that the same spatial point can be used for both binaural devices 110, 112. Furthermore, for a given binaural device, any point in space can be selected as the spatial point from which to virtualize the generated acoustic signals. (The selected point in space can be a moving point in space, e.g., to simulate an audio-generating object in motion.) For example, left, right, or center channel audio signals can be simulated as though they were generated at a location proximate the perimeter speakers 102. Furthermore, the realism of the simulated sound may be enhanced by adding additional virtual sound sources at positions within the environment, i.e., vehicle cabin 100, to simulate the effects of sound generated at the virtual sound source location being reflected off of acoustically reflective surfaces and back to the listener. Specifically, for every virtual sound source generated within the environment, additional virtual sound sources can be generated and placed at various positions to simulate a first order and a second order reflection of sound corresponding to sound propagating from the first virtual sound source and acoustically reflecting off of a surface and propagating back to the listener's ears (first order reflection), and sound propagating from the first virtual sound source and acoustically reflecting off a first surface and a second surface and propagating back to the listener's ears (second order reflection). Methods of implementing HRTFs and virtual reflections to create spatialized audio are discussed in greater detail in U.S. Pat. Pub. US2020/0037097A1 titled “Systems and methods for sound source virtualization,” the entirety of which is incorporated by reference herein. In an example, the virtual sound source can be located outside the vehicle. Likewise, the first order reflections and second order reflections need not be calculated for the actual surfaces within the vehicle, but rather than can be calculated for virtual surfaces outside the vehicle, to for example, create the impression that the user is in a larger area than the cabin, or at least to optimize the reverb and quality of the sound for an environment that is better than the cabin of the vehicle.


Controller 504 is otherwise configured in the manner of controller 104 described in connection with FIGS. 1A and 1B, which is to say that the spatialized acoustic signals 114, 116 can be augmented (e.g., in a time-aligned manner), with bass content produced by perimeter speakers 102. For example, perimeter speakers 102 can be utilized to produce the bass content of first content signal u1, the upper range content of which is produced by binaural device 110 as a spatialized acoustic signal, perceived by the user at seating position P1 to originate at spatial position SP1. Although the bass content produced by perimeter speakers 102 in first listening zone 106 may not be a stereo signal, the user seated at seating position P1 may still perceive the first content signal u1 as originating from spatial position SP1. Likewise, perimeter speakers can augment the bass content of the second content signal u2—the upper range of which being produced by binaural device 112 as a spatial acoustic signal—in the second listening zone. The user at seating position P2 will perceive the second content signal u2 as originating as spatial position SP2 at the second listening zone with the bass content provided as a mono acoustic signal from perimeter speakers 102.


Although two binaural devices 110, 112 are shown in FIG. 5, it should be understood that only a single spatialized binaural signal (e.g., binaural signal b1) can be provided to one binaural device. Furthermore, it is not necessary that each binaural device provide a spatialized acoustic signal; rather one binaural device (e.g., binaural device 110) can provide a spatialized acoustic signal while another (e.g., binaural device 112) can provide a non-spatialized acoustic signal. Furthermore, as mentioned above, each binaural device can receive the same binaural signal such that each user hears the same content, the bass content of which is augmented by the perimeter speakers 102 (which does not necessarily have to be produced in separate listening zones). Further, the example of FIG. 5 can be extended to any number of listening zones and any number of binaural devices.


Controller 504 can further implement an upmixer, which receives for example, left and right program content signals and generates left, right, center, etc. channels within the vehicle. The spatialized audio, rendered by binaural devices (e.g., binaural devices 110, 112) can be leveraged to enhance the user's perception of the source of these channels. Thus, in effect, multiple virtual sound sources can be selected to accurately create impressions of left, right, center, etc., audio channels.



FIG. 6 depicts a flowchart for a method 600 of providing augmented audio to users in a vehicle cabin. The steps of method 600 can be carried out by a controller (such as controller 504) in communication with a set of perimeter speakers disposed in a vehicle (such as perimeter speakers 102) and further in communication with a set of binaural devices (such as binaural device 110, 112) disposed at respective seating positions within the vehicle.


At step 602, a content signal is received. The content signal can be received from multiple potential sources such as mobile devices, radio, satellite radio, a cellular connection, etc. The content signal is an audio signal that includes a bass content and an upper range content.


At step 604, a spatial audio signal is output to a binaural device according to a position signal indicative of the position of a user's head in a vehicle, such that the binaural device produces a spatial acoustic signal perceived by the user as originating from a virtual source. The virtual source can be a selected position within the vehicle cabin, such as, in an example, near to the perimeter speakers of vehicle. This can be accomplished by filtering and/or attenuating the audio signal output to the binaural device according to a plurality of head-related transfer functions (HRTFs) which adjust acoustic signals to simulate sound from the virtual source (e.g., spatial point SP1, SP2). As the signals are binaural, i.e., relate to both of the listener's ears, the system can utilize one or more HRTFs to simulate sound specific to various locations around the listener. It should be appreciated that the particular left and right HRTFs used can be chosen based on a given combination of azimuth angle and elevation detected between the relative position of the user's left and right ears and the respective spatial position. More specifically, a plurality of HRTFs can be stored in memory and be retrieved and implemented according to the detected position of the user's left and right ears and selected spatial position.


The user's head position can be determined according to the output of a headtracking device (such as headtracking device 506, 508), which can be comprised of, for example, a time-of-flight sensor, a LIDAR device, multiple two-dimensional cameras, wearable-mounted inertial motion units, proximity sensors, or a combination of these components. In addition, other suitable devices are contemplated. The output of the headtracking device can be processed through a dedicated controller (e.g., controller 510) which can implement software or a neural network trained to detect the position of the user's head. An example in which an inertial measurement unit is used is described in more detail in connection with connection with FIGS. 8-12, below.


At step 606, the perimeter speakers are driven such that the bass content of the content signal is produced in the cabin. In this way, the spatial acoustic signal produced by the binaural device is augmented by the perimeter speakers in the vehicle cabin. Detecting the position of a user's head can comprise detecting any part of the user, or of a wearable worn by the user, from which the respective positions of the user's ears or the position of wearable worn by the user can be derived, including detecting the position of the user's ears directly or the position of the wearable directly.


While method 600 describes a method for augmenting a spatial acoustic signal provided by a single binaural device, method 600 can be extended to augmenting the multiple content signals provided by multiple binaural devices by arraying the perimeter speakers to produce the bass content of respective content signals in different listening zones throughout the cabin. The steps of such a method are described in method 400 and in connection with FIGS. 1A and 1B.



FIG. 8A depicts an example in which a user orientation sensor is used for head tracking. User orientation sensor 802 is disposed on (which can include in) a wearable worn on the user's head and outputs a user orientation signal m1. As used in this disclosure, an orientation sensor is comprised of a sensor or a plurality of sensors suitable for detecting the orientation of a body. The output signal of an orientation sensor can represent orientation directly (e.g., as changes in pitch, roll, or yaw of the body) or can contain other data from which orientation can be derived, such as accelerations in one or more directions, specific force, or angular rate (other suitable types of data is further contemplated). In one example, an orientation sensor can be an inertial measurement unit. An inertial measurement unit can include, for example, accelerometers, gyroscopes, and/or magnetometers and can output a signal representing orientation directly or other measurements, such as specific force and angular rate. Alternatively, sensors such as accelerometers, gyroscopes, and/or magnetometers can be used, apart from an inertial measurement unit, to determine the orientation of the user. User orientation signal m1 can be used by controllers 510, 504 to provide spatial audio signal b1 to the binaural device. (As described above, in various alternative examples, a single controller, such as controller 504, can be used to both process the position signal according to the user orientation sensor 802 and output the spatial signal to the binaural device. Other architectures and combinations of controllers are contemplated herein.)


The wearable on which user orientation sensor 802 is disposed can be binaural device 110 (i.e., when binaural device 110 is a wearable, such as shown, for example, FIGS. 2 and 3). However, in alternative examples, binaural device 110 may be disposed elsewhere, such as in the headrest, and user orientation sensor 802 can otherwise be worn on the user's head.


However, while user orientation sensor 802 will translate motion from the user's head into an orientation signal, motion introduced by the vehicle, e.g., resulting from turns or bumps, will likewise be picked up by user orientation sensor 802. To distinguish between the motion of the vehicle and the user's head, a separate signal, vehicle orientation signal m2, representative of the orientation of the vehicle, can be employed to isolate changes in orientation introduced by motion of the user's head from that of the vehicle. Stated differently, changes in orientation common to vehicle orientation signal m2 and to user orientation signal m2 are attributable to the vehicle, and thus the motion of the user's head can be isolated by finding the difference between the user orientation signal m1 and the vehicle orientation signal m2. Controller 510 can thus determine the orientation of the user's head relative to the vehicle (i.e., the vehicle acts as the frame of reference for the motion of the head) by finding the difference between the user orientation signal m1 and the vehicle orientation signal m2. As described in this disclosure, the difference can include three-dimensional differences, and can found through subtraction, including vector subtraction or its equivalents, although other methods of finding a difference, such as through a machine learning algorithm, can be used.


Any suitable input representative of the orientation (or changes in orientation) of the vehicle can be used. For example, as shown in FIG. 8B vehicle orientation sensor 804 can be disposed on (which can include in) the vehicle, outputting vehicle orientation signal m2. For optimal performance, vehicle orientation sensor 804 must be fixed to the vehicle so that changes in vehicle orientation are picked up by vehicle orientation sensor 804. In one example, vehicle orientation sensor 804 can be attached to a location inside of the vehicle or an outer surface of the vehicle. Vehicle orientation sensor 804 can be fixed to the vehicle during manufacture, or, alternatively, can retrofitted to the vehicle by a user. For example, vehicle orientation sensor 804 can be disposed, e.g., within a puck or mobile device, that the user brings into the vehicle and attaches to a fixed location, such as the dashboard or center console.


Other suitable sources of vehicle orientation signal m2 include an input representative of vehicle parameters, such as the speed, acceleration, steer angle, etc., which can likewise be used to determine change in orientation of the vehicle. In one example, as shown in FIG. 8C, these parameters can be received from the vehicle control unit 808 by controller 510 and used to calculate the change in orientation of the vehicle, which can then be subtracted (or otherwise removed, such as through a machine learning algorithm) from the user orientation signal m1 to isolate the orientation of the user's head. Other potentially suitable inputs include navigation data, e.g., as determined from the GPS or cellular signals, or camera data (e.g., as used in a self-driving vehicle) can be used as inputs to determine orientation of the vehicle.


Where two orientation sensors are used to track the motion of the user's head and the vehicle (e.g., as shown in FIG. 8B), small errors in measuring orientation, in the aggregate, can manifest in drift between the measured orientations of the user's head and the vehicle. This drift results error in the measured orientation of the user's head, meaning that the user will perceive virtual sound source (e.g., SP1) as being incorrectly positioned and drifting with respect to the user's head.


For the purposes of explanation, FIGS. 9A-9C depict a simplified example of the relative drift between the measured yaw of the user's head and the of the vehicle. FIG. 9A depicts a properly calibrated and measured orientation of the user and the vehicle. In this example, both the measured orientations (m1, m2) are directed in the same direction as the orientation of the vehicle and the user (depicted as dotted lines). In FIG. 9B, the orientation represented by user orientation signal m1 and vehicle orientation signal m2 are both approximately 45° off the correct orientation. Although the measured orientation does not match the actual orientation, since both measured orientations have the same error, the spatial audio signal will be rendered correctly. This is because the vehicle is used as the frame of reference for the spatialized audio. In FIG. 9C, however, the orientation represented by user orientation signal m1 and vehicle orientation signal m2 have drifted relative to each other. In this example, since virtual sound source is positioned at a point relative to the vehicle, the user will perceive the virtual sound source as incorrectly located in space (i.e., in the wrong location in the vehicle).


To correct for relative drift between the user orientation signal m1 and vehicle orientation signal m2, a separate error sensor can be periodically sampled to locate the orientation of the user's head relative to the vehicle. For example, if controller 510, according to orientation sensors 802, 804, determines that the user's head is angled relative to the orientation of the vehicle (e.g., as though the user is looking out the window) but the error sensor determines that the user's head is not angled relative to the orientation of the vehicle (e.g., the user is looking straight ahead), controller 510 can correct for the drift as measured by the orientation sensors 802, 804 by removing the drift (e.g., through subtraction or through other methods, such as machine learning).


Because the relative drift between user orientation signal m1 and vehicle orientation signal m2 can take some time to accumulate, the error sensor need not be sampled as quickly as user orientation signal m1 and vehicle orientation signal m2 are sampled. For example, if user orientation signal m1 and vehicle orientation signal m2 are sampled every millisecond, the error sensor can be sampled every second. (These sampling rates are only provided as examples and other suitable sampling rates can be used.) For example, as shown in FIG. 8B, headtracking device 506 such as a time-of-flight sensor, a LIDAR device, or one or more cameras can be used as error sensors. Using these types of sensors to determine the orientation of the user's head are computationally expensive, and so by only using these sensors to correct for drift, they can be sampled at a slower rate, saving computational resources.


In another example, the error sensor can be at least one microphone located on the wearable. In one example, the error sensor can be two or more microphones disposed on opposing sides of the user's head. The orientation of the wearable can be calculated by measuring a delay between the receipt of an acoustic signal at each of the microphones. An example of this is shown in FIGS. 10A-10B. As shown, a wearable—here, a pair of Bose Frames 200—includes a microphone 1002a and 1002b, respectively disposed on opposing temples. An acoustic signal is emitted from a speaker, such as perimeter speaker 102a (although any suitable speaker can be used). In the orientation of frames 200 depicted in FIG. 10A, microphone 1002a is approximately distance d1 from speaker 102a, while microphone 1002b is approximately distance 102b is distance d2 from speaker 102a. As a result, microphone 1002a will receive the acoustic signal produced by speaker 102a before microphone 1002b receives the acoustic signal from speaker 102a. Stated differently, there will be some delay between the microphone 1002a and microphone 1002b receiving the same acoustic signal produced by speaker 102a.


If the user's head turns, such as shown in FIG. 10B, the distance between microphones 1002a, 1002b and speaker 102a changes. In this example, the distance from each microphone 1002a, 1002b to speaker 102a becomes approximately the same, d3, and will thus receive the same signal at approximately the same time. The lack delay indicates that the user is facing away from (or toward perimeter) speaker 102a. The delay between the receipt of the acoustic signal thus becomes a proxy for measuring the distance between each microphone 1002a, 1002b and speaker 102a.


By monitoring the delay (including the lack of delay) from receipt of a common acoustic signal at each microphone disposed on opposing sides of the user's head, the orientation of the user's head relative to the speaker can be determined. Controller 510, for example, can thus receive the output of microphones 1002a, 1002b and implement a lookup table, or perform a calculation, to translate the delay into an orientation of the user's head relative to the vehicle (as the speaker is in a known position within the vehicle). The time at which the same signal is received can be determined, for example, by performing a measure a similarity calculation (e.g., a cross-correlation) between samples of microphone 1002a and 1002b. This method for detecting orientation is particularly useful for determining the yaw of the user's head relative to the vehicle. Like the other examples of detecting the orientation of the user's head, this example can be used to correct for drift of the relative orientations detected by orientation sensors 802, 804.


In an alternative example, rather than multiple microphones, a single microphone can be employed on the wearable for detection of the orientation of the user's head. By comparing the arrival times of acoustic signals from two or more acoustic sources (e.g., perimeter speakers 102) disposed in known locations, the orientation of the microphones relative to the known speakers can be determined. If the arrival time of a first acoustic signal is earlier than a second acoustic signal, it can be determined the source producing the first acoustic signal is disposed nearer to the microphone than the source producing the second acoustic signal, provided that the acoustic signals are produced at the same time. (Alternatively, the acoustic signals need not be produced at the same time, if the times that the acoustic signals are produced, or the delay between the production of the acoustic signals, are known.)


These and further examples of detecting the position or orientation of a wearable using one or more microphones are further described in US 2022/0082688, titled “Methods and systems for determining position and orientation of a device using acoustic beacons,” which is hereby incorporated by reference in its entirety.


In another example, rather than employing an error sensor, differences in outputs by discrete sensors located within each user orientation sensor 802, 804 can be used to determine drift between the orientation sensors. An example of this is described in connection with FIGS. 11A and 11B, which depict a simplified representation of the individual output signals of a plurality of accelerometers disposed mutually orthogonal within each orientation sensor. (Such an arrangement is common in inertial measurement units, which, as described above, is an example of a suitable orientation sensor.) More particularly, FIG. 11A depicts the outputs of three accelerometers within user orientation sensor 802, each respectively positioned to detect accelerations in particular direction, as denoted by the three-dimensional axis. Thus, plot 1102 depicts the output of an accelerometer disposed to detect accelerations on the z-axis, plot 1104 depicts the output of an accelerometer disposed to detect accelerations on the x-axis, and plot 1106 depicts the output of an accelerometer disposed to detect accelerations on the y-axis. Likewise, FIG. 11B depicts the output of the accelerometers disposed within vehicle orientation sensor 804. Thus, plot 1108 depicts the output of an accelerometer disposed to detect accelerations on the z-axis, plot 1110 depicts the output of an accelerometer disposed to detect accelerations on the x-axis, and plot 1112 depicts the output of an accelerometer disposed to detect accelerations on the y-axis.


Comparing which accelerometer of each orientation sensor detects a common signal, e.g., an acceleration resulting from motion, can reveal the relative orientation of user orientation sensor 802 to vehicle orientation sensor 804. For example, as shown in FIGS. 11A and 11B, the same measured acceleration, caused, e.g., by the vehicle hitting a bump in the road, is detected by the z-axis accelerometer, plot 1102, in user orientation sensor 802 but detected by the x-axis accelerometer, plot 1110, of vehicle orientation sensor 804. It can thus be determined that the z-axis accelerometer of user orientation sensor 802 is pointed in the same direction as the x-axis accelerometer of vehicle orientation sensor 804. Thus, controller 510 can determine the relative orientations of each user orientation sensor 802 and 804 by comparing the similarities in outputs of each accelerometer. If two accelerometers pick up the same acceleration, it can be assumed that they are pointing in the same direction.


The similarity between the outputs of each accelerometer can be determined in any suitable fashion. For example, a cross-correlation can be found between each accelerometer. Thus, a cross-correlation can be found between the x-axis accelerometer output of user orientation sensor 802 and each of the outputs of x-axis, y-axis, and z-axis accelerometers of the vehicle orientation sensor 804. This can be repeated for the y-axis accelerometer and z-axis accelerometer of the user orientation sensor 802 to find a complete mapping of measures of similarities between each accelerometer. It is unlikely that, in most instances, that one accelerometer will be perfectly aligned with anther accelerometer, and thus the relative orientation of user orientation sensor 802 to vehicle orientation sensor 804 can be determined by comparing the degree of similarity each accelerometer output has to each other accelerometer output. By finding the degree of similarity between each accelerometer output, the relative orientation of user orientation sensor 802 to vehicle orientation sensor 804 can be found by controller 510. In practice, a lookup table can be used to compare measures of similarities between each accelerometer to find relative orientation.


While the above examples used an accelerometer output, it should be understood that any kind of sensor output can be compared in this manner. Inertial measurement units, for example, typically also contain three orthogonally positioned gyroscopes, and three orthogonally positioned magnetometers. The similarities in outputs of these sensors can likewise be compared to determine the relative orientations of the user orientation sensor 802 and the vehicle orientation sensor 804. In addition, the outputs other types of sensors, besides those within an inertial measurement unit, or of the same types of sensors outside of an inertial measurement unit, can be used in like fashion to determine the relative orientations of the user orientation sensor 802 and the vehicle orientation sensor 804.


Because comparing the similarity in outputs of sensors is not subject to previous measurements, this method of finding the relative orientation of user orientation sensor 802 to vehicle orientation sensor 804 is immune to drift. Accordingly, this method can be used to correct relative drift between orientation sensors 802 and 804. Alternatively, the relative orientation can be found this way exclusively, rather than relying on the orientations output from the orientation sensors 802, 804, but because this is more intensive calculation, it is more suited to correcting error.


It should be understood that multiple orientation sensors (e.g., additional user orientation sensor 806, as shown in FIG. 8B) can be used to track the orientation of additional passengers within the vehicle (e.g., passenger seated at position P2). Further, as described above, perimeter speakers 102 can be employed to create separate bass zones, such that users seated in separate seating positions can experience spatial audio using binaural devices with bass augmented by the perimeter speakers.



FIG. 12 depicts a flowchart for a method 1200 of providing spatialized audio to a binaural device within a vehicle according to the orientation of a user's head relative to a vehicle. The steps of method 1200 can be carried out by a controller (such as controller 504) or a combination of controllers (such as controller 510 and 504) in communication with a set of perimeter speakers disposed in a vehicle (such as perimeter speakers 102) and further in communication with a set of binaural devices (such as binaural device 110, 112) disposed at respective seating positions within the vehicle.


At step 1202, a user orientation signal output is received from a user orientation sensor disposed on a wearable that, during use, moves with a first user's head. In various examples, the wearable can be a binaural device that is worn by a user (e.g., as shown in FIGS. 2 and 3). In alternative examples, however, the wearable can be separate from the binaural device (which can be located apart from the user, such as within the headrest).


At step 1204, a vehicle orientation signal is received. The vehicle orientation can be received from a second orientation sensor. The vehicle orientation sensor can be disposed within vehicle (e.g., during manufacture) or can be brought within the vehicle, such as in mobile device or puck, and fixed to the vehicle. Alternatively, a separate source of a vehicle orientation can be used, such as an input received from a vehicle control unit with data representative of the vehicle parameters like speed, acceleration, and steering angle. It is conceivable, however, that other inputs, such as navigation data, or camera data used for a self-driving vehicle, could be employed separately or in combination with other methods for determining the orientation of the vehicle.


At step 1206, an orientation of the user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the user orientation signal. To isolate the motion of the user's head relative to the vehicle, the difference can be found between the user orientation signal and the vehicle orientation signal. The difference can be found through, e.g., subtraction, including subtraction of vectors, although other methods of finding the difference, such as with a machine learning algorithm, can be used.


At step 1208, a first spatial audio signal is output to a first binaural device, according to the orientation of the user's head relative to the vehicle, such that the first binaural device produces a first spatial acoustic signal perceived by the first user as originating from a first virtual source location within the vehicle cabin. The position signal, which, here, is representative of the orientation of the user's head relative to the vehicle, provides the basis for locating the spatial audio signal within the vehicle such that the user perceives it as fixed relative to the vehicle. (This method can be repeated for any number of user's wearing an orientation sensor. As described above, the various head orientations of separate users can be used to provide spatialized audio for different users located in separate bass zones, throughout the cabin, as created by arrayed perimeter speakers.)


At step 1210, a drift between the user orientation signal and the vehicle orientation signal is corrected according to a detected orientation of the user's head. This can be accomplished by periodically sampling an error sensor, such as a time-of-flight sensor, a LIDAR device, or one or more cameras, to determine the position of the user's head relative to the vehicle. Alternatively, the error sensor can be one more microphones disposed on one side or on opposing sides of a user's head, used to detect the delay of receipt of the same signal or of different acoustic signals produced at known times by acoustic sources disposed in known locations.


In another example, where analysis of common signals between sensors can reveal orientation of one orientation signal relative to another. This can be accomplished by comparing a measure of similarity (e.g., a cross-correlation) between sensors of the user orientation sensor to sensors of the vehicle orientation sensor. Correlation between signals of different sensors generally corresponds to the degree to which each sensors is disposed in a common direction. Thus, orientation of two orientation sensors can be determined through analysis of similarity of the different sensor signals.


The orientation, as determined by the error sensor or through analysis of signals common to the sensor(s) of the orientation sensors, can be used to correct for drift (e.g., through subtraction or through another other methods, such as machine learning) of the detected orientation of the user's head by the user orientation sensor. Because it takes some time for the drift to accumulate, the error sensor can be sampled, or the orientation otherwise separately determined, at rate that is slower than the rate at which the user orientation sensor disposed on the user's head (as well as the vehicle orientation sensor) is typically sampled.


The functionality described herein, or portions thereof, and its various modifications (hereinafter “the functions”) can be implemented, at least in part, via a computer program product, e.g., a computer program tangibly embodied in an information carrier, such as one or more non-transitory machine-readable media or storage device, for execution by, or to control the operation of, one or more data processing apparatus, e.g., a programmable processor, a computer, multiple computers, and/or programmable logic components.


A computer program can be written in any form of programming language, including compiled or interpreted languages, and it can be deployed in any form, including as a stand-alone program or as a module, component, subroutine, or other unit suitable for use in a computing environment. A computer program can be deployed to be executed on one computer or on multiple computers at one site or distributed across multiple sites and interconnected by a network.


Actions associated with implementing all or part of the functions can be performed by one or more programmable processors executing one or more computer programs to perform the functions of the calibration process. All or part of the functions can be implemented as, special purpose logic circuitry, e.g., an FPGA and/or an ASIC (application-specific integrated circuit).


Processors suitable for the execution of a computer program include, by way of example, both general and special purpose microprocessors, and any one or more processors of any kind of digital computer. Generally, a processor will receive instructions and data from a read-only memory or a random access memory or both. Components of a computer include a processor for executing instructions and one or more memory devices for storing instructions and data.


While several inventive embodiments have been described and illustrated herein, those of ordinary skill in the art will readily envision a variety of other means and/or structures for performing the function and/or obtaining the results and/or one or more of the advantages described herein, and each of such variations and/or modifications is deemed to be within the scope of the inventive embodiments described herein. More generally, those skilled in the art will readily appreciate that all parameters, dimensions, materials, and configurations described herein are meant to be exemplary and that the actual parameters, dimensions, materials, and/or configurations will depend upon the specific application or applications for which the inventive teachings is/are used. Those skilled in the art will recognize, or be able to ascertain using no more than routine experimentation, many equivalents to the specific inventive embodiments described herein. It is, therefore, to be understood that the foregoing embodiments are presented by way of example only and that, within the scope of the appended claims and equivalents thereto, inventive embodiments may be practiced otherwise than as specifically described and claimed. Inventive embodiments of the present disclosure are directed to each individual feature, system, article, material, and/or method described herein. In addition, any combination of two or more such features, systems, articles, materials, and/or methods, if such features, systems, articles, materials, and/or methods are not mutually inconsistent, is included within the inventive scope of the present disclosure.

Claims
  • 1. A system for providing spatialized audio in a vehicle, comprising: a vehicle orientation sensor outputting a vehicle orientation signal and being disposed on the vehicle; anda controller configured to receive a user orientation signal output from a user orientation sensor being disposed on a wearable that, during use, moves with a first user's head, wherein the controller is further configured to determine an orientation of the user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the user orientation signal, the controller being further configured to output to a first binaural device, according to the orientation of the user's head relative to the vehicle, a first spatial audio signal, such that the first binaural device produces a first spatial acoustic signal perceived by the user as originating from a first virtual source location within a cabin of the vehicle.
  • 2. The system of claim 1, further comprising an error sensor configured to detect the orientation of the user's head relative to the vehicle and to output an error sensor signal, wherein the controller is further configured to correct a drift between the user orientation signal and the vehicle orientation signal according to the orientation of the user's head detected by the error sensor.
  • 3. The system of claim 2, wherein the error sensor includes at least one of: time-of-flight sensor, a LIDAR device, a camera, or a microphone disposed on the user's head.
  • 4. The system of claim 2, wherein the controller samples the error sensor signal at a rate slower than the rates at which the vehicle orientation signal and the user orientation signal are sampled.
  • 5. The system of claim 1, wherein the vehicle orientation sensor comprises a first plurality of sensors, wherein the user orientation sensor comprises a second plurality of sensors, wherein the controller is further configured to correct a drift between the user orientation signal and the vehicle orientation signal according to a measure of similarity between at least one sensor of the first plurality of sensors and one sensor of the second plurality of sensors.
  • 6. The system of claim 1, wherein the controller comprises a first controller and a second controller, the first controller receiving the vehicle orientation signal and the user orientation signal and determining the orientation of the user's head relative to the vehicle and outputting a position signal to the second controller, the second controller, receiving the position signal, outputting to the first spatial acoustic signal to the first binaural device.
  • 7. The system of claim 1, wherein the wearable is the first binaural device.
  • 8. The system of claim 1, further comprising: a plurality of speakers disposed in a perimeter of a cabin of the vehicle,wherein the first spatial audio signal comprises at least an upper range of a first content signal,wherein the controller is further configured to drive the plurality of speakers with a driving signal such that a first bass content of the first content signal is produced in the cabin.
  • 9. The system of claim 8, wherein the controller is configured to receive a second user orientation signal output from a second user orientation sensor being disposed on a second wearable that, during use, moves with a second user's head, wherein the controller is further configured to determine an orientation of the second user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the second user orientation signal, the controller being further configured to output to a second binaural device, according to the orientation of the second user's head relative to the vehicle, a second spatial audio signal, such that the second binaural device produces a second spatial acoustic signal perceived by the second user as originating from the first or a second virtual source location within a cabin of the vehicle.
  • 10. The system of claim 9, wherein the second spatial audio signal comprises at least an upper range of a second content signal, wherein the controller is further configured to drive the plurality of speakers in accordance with a first array configuration such that the first bass content is produced in a first listening zone within the cabin and in accordance with a second array configuration such that a second bass content of the second content signal produced in a second listening zone within the cabin, wherein in the first listening zone a magnitude of the first bass content is greater than a magnitude of the second bass content and in the second listening zone the magnitude of the second bass content is greater than the magnitude of the first bass content.
  • 11. The system of claim 1, wherein the vehicle orientation sensor is: integrated with the vehicle or brought into and fixed to the vehicle by a user.
  • 12. A method for providing spatialized audio in a vehicle, comprising: receiving a user orientation signal output from a user orientation sensor being disposed on a wearable that, during use, moves with a first user's head;receiving a vehicle orientation signal from a vehicle orientation sensor being disposed on the vehicle;determining an orientation of the user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the user orientation signal; andoutputting to a first binaural device, according to the orientation of the user's head relative to the vehicle, a first spatial audio signal, such that the first binaural device produces a first spatial acoustic signal perceived by the user as originating from a first virtual source location within a cabin of the vehicle.
  • 13. The method of claim 12, further comprising: Receiving an error sensor signal output from an error sensor configured to detect the orientation of the user's head relative to the vehicle; andcorrecting a drift between the user orientation signal and the vehicle orientation signal according to the orientation of the user's head detected by the error sensor.
  • 14. The method of claim 13, wherein the error sensor includes at least one of: a time-of-flight sensor, a LIDAR device, a camera, or a microphone disposed on the user's head.
  • 15. The method of claim 13, wherein the error sensor signal is sampled at a rate slower than the rates at which the vehicle orientation signal and the user orientation signal are sampled.
  • 16. The method of claim 12, further comprising the step of correcting a drift between the user orientation signal and the vehicle orientation signal, wherein the vehicle orientation sensor comprises a first plurality of sensors, wherein the user orientation sensor comprises a second plurality of sensors, wherein the drift is corrected according to a measure of similarity between at least one sensor of the first plurality of sensors and one sensor of the second plurality of sensors.
  • 17. The method of claim 12, wherein the wearable is the first binaural device.
  • 18. The method of claim 12, further comprising: driving a plurality of speakers with a driving signal such that a first bass content of the first spatial audio signal is produced in the cabin.
  • 19. The method of claim 18, further comprising receiving a second user orientation signal output from a second user orientation sensor being disposed on a second wearable that, during use, moves with a second user's head,determining an orientation of the second user's head relative to the vehicle based, at least, on a difference between the vehicle orientation signal and the second user orientation signal,outputting to a second binaural device, according to the orientation of the second user's head relative to the vehicle, a second spatial audio signal, such that the second binaural device produces a second spatial acoustic signal perceived by the second user as originating from the first or a second virtual source location within a cabin of the vehicle.
  • 20. The method of claim 19, further comprising: driving the plurality of speakers in accordance with a first array configuration such that the first bass content is produced in a first listening zone within the cabin and in accordance with a second array configuration such that a second bass content of the second spatial audio signal is produced in a second listening zone within the cabin, wherein in the first listening zone a magnitude of the first bass content is greater than a magnitude of the second bass content and in the second listening zone the magnitude of the second bass content is greater than the magnitude of the first bass content.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims priority to U.S. Provisional Patent Application Ser. No. 63/366,294, filed on Jun. 13, 2022, and titled “Systems and Methods for Providing Augmented Audio,” which application is herein incorporated by reference in its entirety.

Provisional Applications (1)
Number Date Country
63366294 Jun 2022 US