This application is directed, in general, to speakers and, more specifically, to directing sound transmission.
Acoustic transducers are used when converting sound from one form of energy to another form of energy. For example, microphones are used to convert sound to electrical signals (i.e., an acoustic-to-electric transducer). The electrical signals can then be processed (e.g., cleaned-up, amplified) and transmitted to a speaker or speakers (hereinafter referred to as a loudspeaker or loudspeakers). The loudspeakers are then used to convert the processed electrical signals back to sound (i.e., an electric-to-acoustic transducer).
Often, such as in a concert or a speech, the loudspeakers are arranged to provide audio-coverage throughout an area. In other words, the loudspeakers are arranged to propagate sound received from a microphone or microphones throughout a designated area. Therefore, each person in the area is able to hear the transmitted sound.
One aspect provides a directional sound system. In one embodiment, the directional sound system includes: (1) a direction sensor configured to produce data for determining a direction in which attention of a user is directed, (2) a microphone configured to generate output signals indicative of sound received thereat, (3) loudspeakers configured to convert directed sound signals into directed sound and (4) an acoustic processor configured to be coupled to the direction sensor, the microphone, and the loudspeakers, the acoustic processor configured to convert the output signals to the directed sound signals and employ the loudspeakers to transmit the directed sound to a spatial location associated with the direction.
Another aspect provides a method of transmitting sound to a spatial location determined by the gaze of a user. In one embodiment, the method includes: (1) determining a direction of visual attention of a user associated with a spatial location, (2) generating directed sound signals indicative of sound received from a microphone, (3) converting the directed sound signals to directed sound employing loudspeakers having known positions relative to one another and (4) transmitting the directed sound in the direction employing the loudspeakers to provide directed sound at the spatial location.
Still yet another aspect provides a directional communication system. In one embodiment, the directional communication system includes: (1) an eyeglass frame, (2) a direction sensor on the eyeglass frame and configured to provide data indicative of a direction of visual attention of a user wearing the eyeglass frame, (3) a microphone configured to generate output signals indicative of sound received thereat, (4) acoustic transducers arranged in an array and configured to provide output signals indicative of sound received at the microphone and (5) an acoustic processor coupled to the direction sensor, the microphone, and the acoustic transducers, the acoustic processor configured to convert the output signals to directed sound signals and employ the acoustic transducers to transmit directed sound based on the directed sound signals to a spatial location associated with the direction.
Reference is now made to the following descriptions taken in conjunction with the accompanying drawings, in which:
Instead of propagating sound throughout an area, this disclosure addresses how sound can be directed to a spatial location (e.g., a spatial volume). As such, a human speaker can direct the sound of his voice selectively to a spatial location. Thus, a speaker could selectively speak to another person while limiting the ability of other people in the area to hear what is spoken. In some embodiments, the speaker could selectively speak over a considerable distance to another person.
As disclosed herein, a steerable loudspeaker array can be combined with a direction sensor to direct sound. The steerable loudspeaker array may be electronically-steerable or even mechanically-steerable. The user could speak (or whisper) into a microphone, and the sound of his voice can be transmitted selectively by the loudspeaker array towards the point in space, or even points in space, at which the user is looking. This may be performed without requiring special equipment for the party towards whom the sound is directed. The sound may be transmitted to the point in space in stereo.
The direction sensor may be an eye-tracking device such as a non-contact eye-tracker that is based on infrared light reflected from a cornea. Nanosensors may be used to provide a compact eye-tracker that could be built into eye-glass frames. Other types of direction sensors, such as a head tracking device, may also be used.
The loudspeaker array is to be sufficiently large enough (both with respect to spatial extent and the number of loudspeakers) to provide a desired angular resolution for directing the sound. The loudspeaker array may include loudspeakers built into the user's clothing and additional loudspeakers coupled to these loudspeakers to augment the user's array. The additional loudspeakers may be wirelessly linked. The additional loudspeakers may be attached to other users or fixed at various locations.
Processing of the acoustic signals may occur in real-time. Under line-of-sight propagation conditions, delay-and-sum beamforming could be used. Under multipath conditions, a more general filter-and-sum beamformer might be effective. If the user were directing the sound to another human speaker, and if the other user spoke, then reciprocity would aid the beamforming process. In some embodiments, a microphone array can be co-located with a loudspeaker array. The microphone array, for example, may be the array disclosed in U.S. patent application Ser. No. 12/238,346, entitled “SELF-STEERING DIRECTIONAL HEARING AID AND METHOD OF OPERATION THEREOF,” by Thomas L. Marzetta, filed on Sep. 25, 2008, and incorporated herein by reference in its entirety and referred to herein as Marzetta. Instead of a separate array of microphones, an array of acoustic transducers may be used that operate as both microphones and loudspeakers.
In one embodiment, the direction sensor is associated with any portion of the head of the user 100 as a block 110a indicates. This allows the direction sensor to produce a head position signal that is based on the direction in which the head of the user 100 is pointing. In a more specific embodiment, the direction sensor is proximate one or both eyes of the user 100 as a block 110b indicates. This allows the direction sensor to produce an eye position signal based on the direction of the gaze of the user 100. Alternative embodiments locate the direction sensor in other places that still allow the direction sensor to produce a signal based on the direction in which the head or one or both eyes of the user 100 are pointed. A pointing device may also be used with a direction sensor to indicate a spatial location. For example, as represented by block 120b, the user 100 may use a direction sensor with a directional indicator, such as a wand or a laser beam, to associate movements of a hand with a location signal that indicates the spatial location. The directional indicator may wirelessly communicate with a direction sensor to indicate the spatial location based on movements of the directional indicator by the hand of the user. In some embodiments, the directional indicator may be connected to the direction sensor via a wired connection.
The direction sensor may be used to indicate two or more spatial locations based on head positions or gaze points of the user 100. As such, the loudspeakers can be positioned to simultaneously transmit sound to each of the different spatial locations. For example, a portion of the loudspeakers may be positioned to transmit directed sound to one spatial location while other loudspeakers may be positioned to simultaneously transmit the directed sound to another or other spatial locations. Additionally, the size of the spatial location identified by the user 100 may vary based on the head positions or gaze points of the user. For example, the user 100 may indicate that the spatial location is a region by moving his eyes in a circle. Thus, instead of multiple distinct spatial locations for simultaneous transmission, the loudspeakers may be directed to transmit sound to a single, contiguous spatial location that could include multiple people.
The microphone is located proximate the user 100 to receive sound to be transmitted to a spatial location according to the direction sensor. In one embodiment, the microphone is located proximate the mouth of the user 100, as indicated by block 120a, to capture the user's voice for transmission. The microphone may be attached to clothing worn by the user 100 using a clip. In some embodiments, the microphone may be attached to the collar of the clothing (e.g., a shirt, a jacket, a sweater or a poncho). In other embodiments, the microphone may be located proximate the mouth of the user 100 via an arm connected to a headset or eyeglass frame. The microphone may also be located proximate the arm of the user 100 as indicated by a block 120b. For example, the microphone may be clipped to a sleeve of the clothing or attached to a bracelet. As such, the microphone can be placed proximate the mouth of the user when desired by the user.
In one embodiment, the loudspeakers are located within a compartment that is sized such that it can be placed in a shirt pocket of the user 100 as a block 130a indicates. In an alternative embodiment, the loudspeakers are located within a compartment that is sized such that it can be placed in a pants pocket of the user 100 as a block 130b indicates. In another alternative embodiment, the loudspeakers are located proximate the direction sensor, indicated by the block 110a or the block 110b. The aforementioned embodiments are particularly suitable for loudspeakers that are arranged in an array. However, the loudspeakers need not be so arranged. Therefore, in yet another alternative embodiment, the loudspeakers are distributed between or among two or more locations on the user 100, including but not limited to those indicated by the blocks 110a, 110b, 130a, 130b. In still another alternative embodiment, one or more of the loudspeakers are not located on the user 100 (i.e., the loudspeakers are located remotely from the user), but rather around the user 100, perhaps in fixed locations in a room in which the user 100 is located. One of more of the loudspeakers may also be located on other people around the user 100 and wirelessly coupled to other components of the directional sound system.
In one embodiment, the acoustic processor is located within a compartment that is sized such that it can be placed in a shirt pocket of the user 100 as the block 130a indicates. In an alternative embodiment, the acoustic processor is located within a compartment that is sized such that it can be placed in a pants pocket of the user 100 as the block 130b indicates. In another alternative embodiment, the acoustic processor is located proximate the direction sensor, indicated by the block 110a or the block 110b. In yet another alternative embodiment, components of the acoustic processor are distributed between or among two or more locations on the user 100, including but not limited to those indicated by the blocks 110a, 110b, 120a, 120b. In still other embodiments, the acoustic processor is co-located with the direction sensor, with the microphone or one or more of the loudspeakers.
The microphone 141 is configured to provide output signals based on received acoustic signals, called “raw sound” in
The acoustic processor 143 is coupled by wire or wirelessly to the microphone 141 and the loudspeakers 147. The acoustic processor 143 may be a computer including a memory having a series of operating instructions that direct its operation when initialized thereby. The acoustic processor 143 is configured to process and direct the output signals received from the microphone 141 to the loudspeakers 147. The loudspeakers 147 are configured to convert the processed output signals (i.e., directed sound signals) from the acoustic processor 143 into directed sound and transmit the directed sound towards a point in space based on a direction received by the acoustic processor 143 from the direction sensor 145.
The directed sound signals may vary for each particular loudspeaker in order to provide the desired sound at the point in space. For example, the directed sound signals may vary based on a transmitting delay to allow beamforming at the point in space. The directed sound signals may also be transmitted in a higher frequency band and shifted back down to the voice band at a receiver at the point in space. An ultrasonic frequency band, for example, may even be used. Using audio frequency-shifting can provide greater directivity using a smaller array of loudspeakers, and possibly more privacy. To increase privacy even more, the frequency shifting could follow a random hopping pattern. When employing the frequency-shifting, a person receiving the directed sound signal at the point in space would use a special receiver configured to receive the transmitted signal and shift the signal down to base-band.
The directed sound signals may also vary to allow stereo sound at the point in space. To provide stereo sound, the loudspeakers may be divided into left and right loudspeakers with each loudspeaker group receiving different directed sound signals to provide stereo sound at the point in space. Alternatively, the entire array of loudspeakers could be driven simultaneously by the sum of two sets of directed sound signals.
The acoustic processor 143 employs the received direction, the known relative position of the loudspeakers 147 to one another and the orientation of the loudspeakers 147 to direct each loudspeaker of the loudspeakers 147 to transmit the directed sound to the point in space. The loudspeakers 147 are configured to provide the directed sound based on the received acoustic signals (i.e., the raw sound in
The direction sensor 145 is configured to determine the direction by determining where a user's attention is directed. The direction sensor 145 may therefore receive an indication of head direction, an indication of eye direction, or both, as
The directional communication system 150 allows enhanced communication by providing directed sound to a spatial location and receiving enhanced sound from the spatial location. The acoustic transducers 151 are configured to operate as microphones and loudspeakers. The acoustic transducers 151 may be an array such as the loudspeaker array 230 of
The loudspeaker 155 is coupled, wirelessly or by wire, to the acoustic processor 143. The loudspeaker 155 is configured to convert an enhanced sound signal generated by the acoustic processor 143 into enhanced sound as disclosed in Marzetta.
The array of loudspeakers 230 includes loudspeakers 230a, 230b, 230c, 230d, . . . , 230n. The array of loudspeakers 230 may be a one-dimensional (substantially linear) array, a two-dimensional (substantially planar) array, a three-dimensional (volume) array or any other configuration.
Delays, referred to as transmitting delays, may be associated with each loudspeaker of the array of loudspeakers 230 to control when the sound waves are sent. By controlling when the sound waves are sent, the sound waves can arrive at the point of gaze 220 at the same time. Therefore, the sum of the sound waves will be perceived by a user at the point of gaze 220 to provide an enhanced sound. An acoustic processor, such as the acoustic processor 143 of
Angles θ and φ (see
In an alternative embodiment, the orientation of the array of loudspeakers 230 is determined with an auxiliary orientation sensor (not shown), which may take the form of a position sensor, an accelerometer or another conventional or later-discovered orientation-sensing mechanism.
In the embodiment of
In another embodiment, the acoustic processor 143 may initiate the calibration mode to determine the transmitting delays for each of the loudspeakers 230a-1, . . . , 230m-n with respect to the point of gaze by employing one of the loudspeakers 230a-1, . . . , 230m-n to transmit an audio signal to the point of gaze 220. The other remaining loudspeakers may be used as microphones to receive a reflection of the transmitted audio signal. The acoustic processor 143 can then determine the transmitting delays from the reflected audio signal received by the remaining loudspeakers 230a-1, . . . , 230m-n. This process may be repeated for multiple of the loudspeakers 230a-1, . . . , 230m-n. Processing of the received reflected audio signals, such as filtering, may be necessary due to interference from objects.
The calibration mode may cause acoustic energy to emanate from a known location or determine the location of emanating acoustic energy (perhaps with a camera), capturing the acoustic energy with the loudspeakers (being used as microphones) and determining the amount by which the acoustic energy is delayed with respect to each loudspeaker. Correct transmitting delays may thus be determined. This embodiment is particularly advantageous when loudspeaker positions are aperiodic (i.e., irregular), arbitrary, changing or unknown. In additional embodiments, wireless loudspeakers may be employed in lieu of, or in addition to, the loudspeakers 230a-1, . . . , 230m-n.
The signal 510a contains a transient 540a representing acoustic energy received from a first source, a transient 540b representing acoustic energy received from a second source, a transient 540c representing acoustic energy received from a third source, a transient 540d representing acoustic energy received from a fourth source and a transient 540e representing acoustic energy received from a fifth source.
The signal 510b also contains transients representing acoustic energy emanating from the first, second, third, fourth and fifth sources (the last of which occurring too late to fall within the temporal scope of
Although
One embodiment of the acoustic processor takes advantage of this phenomenon by delaying output signals to be transmitted by each of the acoustic transducers 230a-1, 230a-2, 230a-3 according to the determined relative time delay. The transmitting delay for each of the acoustic transducers 230a-1, 230a-2, 230a-3 is based on the output signal received from the direction sensor, namely an indication of the angle θ, upon which the delay is based.
The following equation relates the delay to the horizontal and vertical pitches and of the microphone relay:
where d is the delay, integer multiples of which the acoustic processor applies to the output signal of each microphone in the array, φ is the angle between the projection of the line 250 of
In
The example of
The example of
An alternative embodiment to that shown in
Those skilled in the art to which this application relates will appreciate that other and further additions, deletions, substitutions and modifications may be made to the described embodiments.