The present technology generally relates to virtual auditory display devices, and more particularly to virtual auditory display devices that output virtual auditory display sound generated by applying virtual auditory display filters to audio signals.
Headphones and earbuds may transmit sound. Similarly, headsets may transmit sound. Headsets may also capture sound using microphones.
In some aspects, the techniques described herein relate to a system including: a first ear-worn device; and a second ear-worn device, wherein the first ear-worn device and the second ear-worn device each include: an acoustic package including: a housing including a first housing portion having a first partial generally capsule shape and a second housing portion having a second partial generally capsule shape; a first set of electrical contacts; and one or more speakers positioned within the housing, the one or more speakers configured to emit sound based on signals received by the first set of electrical contacts; an ear interface removably coupleable to the acoustic package, the ear interface including a proximal portion and a distal portion, a first opening at the proximal portion, a first cavity and a second cavity extending away from the first opening, the first cavity having a third partial generally capsule shape generally matching the first partial generally capsule shape, the second cavity having a fourth partial generally capsule shape generally matching the second partial generally capsule shape, and a second opening at the distal portion through which the sound emitted by the one or more speakers may pass; and an electronics package removably coupleable to the acoustic package, the electronics package including a second set of electrical contacts configured to connect with the first set of electrical contacts, and electronics configured to receive audio signals, generate the signals based on the audio signals, and provide the signals to the first set of electrical contacts via the second set of electrical contacts.
In some aspects, the techniques described herein relate to a system wherein: the housing of the acoustic package further includes a third housing portion having a first generally cylindrical shape, the acoustic package further includes a snout including a snout proximal portion configured to be positioned at least partially within the third housing portion and a snout distal portion, the snout having a third opening at the snout proximal portion, a fourth opening at the snout distal portion, and a snout passage therebetween such that the sound emitted by the one or more speakers may pass through the third opening, the snout passage, and the fourth opening, and the ear interface further includes a passage extending from the first cavity to the second opening at the distal portion, a portion of the passage having a second generally cylindrical shape generally matching the first generally cylindrical shape.
In some aspects, the techniques described herein relate to a system wherein the ear interface further includes a pressure-equalization vent having a fifth opening at one of the first cavity, the second cavity, and the passage, a sixth opening at an exterior of the ear interface, and a vent passage between the fifth opening and the sixth opening.
In some aspects, the techniques described herein relate to a system wherein the acoustic package further includes a cap coupled to the housing and a pressure-equalization vent in the cap, the pressure-equalization vent including one or more layers of acoustic mesh.
In some aspects, the techniques described herein relate to a system wherein the one or more layers of acoustic mesh have a rayl value of approximately 1800.
In some aspects, the techniques described herein relate to a system, further including a cable including a first connector configured to connect to another device, a first cable portion attached to the electronics package of the first ear-worn device, and a second cable portion attached to the electronics package of the second ear-worn device, wherein the cable is configured to transmit a first audio signal to the electronics of the electronics package of the first ear-worn device and a second audio signal to the electronics of the electronics package of the second ear-worn device.
In some aspects, the techniques described herein relate to a system wherein the electronics of the electronics package include a wireless communication component configured to wirelessly receive the audio signals.
In some aspects, the techniques described herein relate to a system wherein: the acoustic package further includes a first magnet coupled to the housing, the electronics package further includes a second magnet, and the electronics package is removably magnetically coupleable to the acoustic package by magnetic attractive forces between the first magnet and the second magnet.
In some aspects, the techniques described herein relate to a system wherein the first ear-worn device and the second ear-worn device each further include a collar removably coupleable to the ear interface, the collar extending generally circumferentially around the proximal portion of the ear interface when coupled to the ear interface.
In some aspects, the techniques described herein relate to a system wherein: the first set of electrical contacts include a set of annular electrical contacts, and the second set of electrical contacts include a set of pogo pins configured to contact the set of annular electrical contacts.
In some aspects, the techniques described herein relate to a system wherein the electronics package further includes one or more processors, one or more memories, and one or more components configured to capture head orientation data for a head orientation of a wearer of the first ear-worn device and the second ear-worn device, the one or more memories storing instructions that when executed by the one or more processors cause the one or more processors to determine the head orientation of the wearer based on the head orientation data captured by the one or more components.
In some aspects, the techniques described herein relate to a system wherein: the electronics package of the first ear-worn device further includes one or more first processors and one or more first memories, the electronics package of the second ear-worn device further includes one or more second processors, and the one or more first memories include instructions that when executed by the one or more first processors cause the one or more first processors to control the one or more second processors.
In some aspects, the techniques described herein relate to a system wherein: the acoustic package further includes a microphone configured to capture sounds of a wearer of the first ear-worn device and the second ear-worn device, and the electronics package further includes one or more processors and one or more memories including instructions that when executed by the one or more processors cause the one or more processors to perform active noise cancellation of the sounds of the wearer.
In some aspects, the techniques described herein relate to a system wherein: the electronics package further includes a microphone configured to capture external sounds, one or more processors and one or more memories including instructions that when executed by the one or more processors cause the one or more processors to perform active noise cancellation of the external sounds.
In some aspects, the techniques described herein relate to a system wherein the signals are first signals, the audio signals are first audio signals, the microphone is a first microphone, the active noise cancellation of the external sounds is a first mode of active noise cancellation of the first ear-worn device and the second ear-worn device, and wherein: the electronics package further includes an inertial measurement unit, a magnetometer, multiple second microphones, and cap, at least one of the inertial measurement unit and the magnetometer are configured to detect interactions of a wearer of the first ear-worn device and the second ear-worn device with the cap, the multiple second microphones are configured to capture the external sounds, the one or more memories include further instructions to cause the one or more processors to switch between the first mode and a second mode of transparency of the first ear-worn device and the second ear-worn device, and when the first ear-worn device and the second ear-worn device are in the second mode, to generate second audio signals based on the external sounds captured by the multiple second microphones, generate second signals based on the second audio signals, and provide the second signals to the one or more speakers, and the one or more speakers are further configured to emit sound based on the second signals.
In some aspects, the techniques described herein relate to a device including: an acoustic package including: a housing including a first housing portion having a first partial generally capsule shape and a second housing portion having a second partial generally capsule shape; and one or more speakers positioned within the housing, the one or more speakers configured to emit sound based on a signal received by the acoustic package; an ear interface including a proximal portion and a distal portion, a first opening at the proximal portion, a first cavity and a second cavity extending away from the first opening, the first cavity having a third partial generally capsule shape generally matching the first partial generally capsule shape, the second cavity having a fourth partial generally capsule shape generally matching the second partial generally capsule shape, and a second opening at the distal portion through which the sound emitted by the one or more speakers may pass; and an electronics package removably coupleable to the acoustic package, the electronics package including electronics configured to receive an audio signal, generate the signal based on the audio signal, and provide the signal to the acoustic package.
In some aspects, the techniques described herein relate to a device wherein: the housing of the acoustic package further includes a third housing portion having a first generally cylindrical shape, the acoustic package further includes a snout including a snout proximal portion configured to be positioned at least partially within the third housing portion and a snout distal portion, the snout having a third opening at the snout proximal portion, a fourth opening at the snout distal portion, and a snout passage therebetween such that the sound emitted by the one or more speakers may pass through the third opening, the snout passage, and the fourth opening, and the ear interface further includes a passage extending from the first cavity to the second opening at the distal portion, a portion of the passage having a second generally cylindrical shape generally matching the first generally cylindrical shape.
In some aspects, the techniques described herein relate to a device wherein the ear interface further includes a pressure-equalization vent.
In some aspects, the techniques described herein relate to a device wherein the acoustic package further includes a pressure-equalization vent, the pressure-equalization vent including one or more layers of acoustic mesh.
In some aspects, the techniques described herein relate to a device wherein the one or more layers of acoustic mesh have a rayl value of approximately 1800.
In some aspects, the techniques described herein relate to a device wherein: the acoustic package further includes a first magnet coupled to the housing, the electronics package further includes a second magnet, and the electronics package is removably magnetically coupleable to the acoustic package by magnetic attraction between the first magnet and the second magnet.
In some aspects, the techniques described herein relate to a device, wherein the electronics package further includes one or more processors, one or more memories, and one or more components configured to capture head orientation data for a head orientation of a wearer of the device, the one or more memories storing instructions that when executed by the one or more processors cause the one or more processors to determine the head orientation of the wearer based on the head orientation data captured by the one or more components.
Throughout the drawings, like reference numerals will be understood to refer to like parts, components, and structures.
Described herein are virtual auditory display devices. A virtual auditory display device may include a first ear-worn device and a second ear-worn device that a person wears on the person's ears. The virtual auditory display device may receive audio signals from another device and generate virtual auditory display sound based on the audio signals. Virtual auditory display sound may refer to sounds that are capable of being perceived by the wearer of the virtual auditory display device as coming from any point in space surrounding the wearer. The virtual auditory display device may thus provide an immersive sound experience for the wearer.
Virtual auditory display devices may have a wide range of uses. For example, a music creator may use a virtual auditory display device to listen to music the music creator has produced. The music creator may then remix or modify the music based on their hearing the music rendered as virtual auditory display sound. For example, the music creator may move the location of certain sounds, emphasize certain sounds, deemphasize certain sounds, and the like. In this fashion, the music creator may utilize the virtual auditory display device as part of an iterative process of creating music until the music creator achieves the desired effect for the music.
Another use may be by music aficionados who may utilize a virtual auditory display device to provide a listening experience that reinvigorates the music that they love. Another use may be for users who play video games. The virtual auditory display devices may allow the users to hear sounds emanating from locations that are not shown on their displays, thereby improving the users' awareness.
Another group of example use cases relate to military, non-military (for example, first responders such as police and firefighters) and/or other organizational applications. Virtual auditory display devices may be used to provide hyper-realistic virtual audio environments that facilitate virtual training for military and/or non-military personnel. For military personnel, virtual auditory display devices and associated devices may provide enhanced hearing, communications and hyper-situational awareness in combat and training. Other uses are described herein, and still other uses will be apparent.
The second ear-worn device 102b includes a second electronics package 104b, a second ear interface 106b, and a second acoustic package 108b positioned within the second ear interface 106b and attached to the second electronics package 104b. The cable 110 includes a connector 116, a cable connector portion 114, a junction 118, a first cable portion 112a connected to the junction 118 and the first electronics package 104a, and a second cable portion 112b connected to the junction 118 and the second electronics package 104b.
As described in more detail herein, the first ear interface 106a is structured to be placed in a left ear of a wearer of the virtual auditory display device 100 and the second ear interface 106b is structured to be placed in a right ear of the wearer. The first ear interface 106a and the second ear interface 106b may be custom made for the wearer's ears so as to provide a generally acoustically sealed fit for the wearer's ears.
Each of the first electronics package 104a and the second electronics package 104b includes electronics components for generating audio signals based on digital audio signals received via the cable 110 from an external device, such as a phone or a computer, to which the virtual auditory display device 100 is connected. Each of the first acoustic package 108a and the second acoustic package 108b includes one or more analog components, such as one or more speakers, that are configured to emit sound based on the audio signals received from the first electronics package 104a and the second electronics package 104b respectively. The sound travels through the first ear interface 106a and the second ear interface 106b and into the ear canals of the wearers.
The wearer may remove the first electronics package 104a and the second electronics package 104b from the first acoustic package 108a and the second acoustic package 108b and insert the first ear interface 106a into the wearer's left ear and the second ear interface 106b into the wearer's right ear. The wearer may then connect the first electronics package 104a to the first acoustic package 108a and the second electronics package 104b to the second acoustic package 108b.
Although the first ear interface 106a and the first acoustic package 108a are for the left ear of the wearer and the second ear interface 106b and the second acoustic package 108b are for the right ear of the wearer, the first electronics package 104a may be connected to either the first acoustic package 108a or the second acoustic package 108b. Similarly, the second electronics package 104b may be connected to either the first acoustic package 108a or the second acoustic package.
The user may connect the virtual auditory display device 100 to an external device such as a phone or laptop or desktop computer (not illustrated in
The first electronics package 104a includes a microphone cover 318 including multiple perforations 316. The first electronics package 104a further includes a magnet 306 positioned inward relative to the microphone cover 318. The microphone cover 318 and the magnet 306 form a generally cylindrical recess 302 having a generally planar surface 304. A second set of electrical contacts 308 extend outwards from the generally planar surface 304. Each electrical contact of the second set of electrical contacts 308 is configured to connect with a separate electrical contact of the first set of electrical contacts 312 so that power and/or data may pass between the electronics package 104a and the acoustics package 108a. In some embodiments, the second set of electrical contacts 308 includes a set of pogo pins. In some embodiments, there are seven pogo pins in the set, and each pogo pin is arranged on the generally planar surface 304 such that the pogo pin contacts a separate annular electrical contact.
The generally cylindrical recess 302 has the same general shape as the magnet 314 and the connector 310 of the second acoustic package 108b. The first electronics package 104a may thus removably magnetically couple to the first acoustic package 108a. The first electronics package 104a is removably magnetically coupleable to the first acoustic package 108a due to attractive magnetic forces between the magnet 314 and the magnet 306. Similarly, the second electronics package 104b of the second ear-worn device 102b is also removably magnetically coupleable to the second acoustic package 108b due to magnetic attractive forces between corresponding magnets.
The first ear interface 106a includes a proximal portion 340, an upper portion 342, and a distal portion 344. When first ear-worn device 102a is worn by a wearer, the distal portion 344 is positioned in the left ear canal of the wearer and the upper portion 342 is positioned generally proximate to the left ear concha and generally between the left ear antihelix and helical crus of the wearer. The proximal portion 340 is positioned generally proximate to the left ear antihelix, antitragus, and tragus of the wearer.
The first ear interface 106a also includes a first opening and multiple cavities (not illustrated in
The first ear interface 106a and the second ear interface 106b (not illustrated in
The first electronics package 104a further includes a cable printed circuit board 404. The first cable portion 112a includes multiple wires 406 that are attached to the cable printed circuit board 404. In some embodiments, the first cable portion 112a includes eleven wires. Several of the multiple wires 406 may be for carrying power and/or data from an external device to which the connector 116 is connected to, such as a phone or a desktop or laptop computer. Several of the multiple wires 406 may be for carrying data to and from the first electronics package 104a.
The first electronics package 104a further includes a printed circuit board 408. The printed circuit board 408 includes multiple electronics components such as a microcontroller, memory, an inertial measurement unit (IMU)-based sensor system (which may be referred to as an IMU), a magnetometer, codecs with audio digital signal processors (DSPs), multiple microphones, and the second set of electrical contacts 308.
In some embodiments, the printed circuit board 408 includes nine microphones. The nine microphones on the printed circuit board 408 may be utilized for different purposes. In some embodiments, eight of the nine microphones are digital and may be utilized to create a transparency mode by capturing external noises that are processed by the first electronics package 104a and output by first acoustic package 108a. In some embodiments, one microphone is a high signal-to-noise ratio analog microphone that may be utilized for feedforward active noise cancellation.
The first electronics package 104a also includes a first pressure-sensitive adhesive layer 410, an electrical connector spacer 412, the magnet 306, and a first pressure-sensitive adhesive layer 416.
The first electronics package 104a also includes a microphone manifold 418 having an opening 426 with a continuously increasing radius that may reduce or eliminate a Helmholtz resonance. The first electronics package 104a further includes a microphone cover adhesive layer 420 and a microphone cover 422. The microphone cover 422 includes numerous perforations 424 to allow sounds to pass through and be captured by the multiple microphones on the printed circuit board 408. The microphone cover 422 may be made of any suitable material, such as stainless steel, and the numerous perforations 424 may be created by chemical etching. In some embodiments a perforation has a diameter of approximately 150 microns.
The components of the first electronics package 104a may be attached or coupled using any suitable means, such as adhesives, mechanical fasteners, ultrasonic welding, and the like. Although not depicted in
The housing 530 also includes a third housing portion 528 having a generally cylindrical shape. Positioned within the housing 530 are various components including one or more speakers, such as a driver and a balanced armature. The housing 530 may be made of any suitable material, such as polycarbonate, and may be transparent, translucent, or opaque.
The first acoustic package 108a further includes the magnet 314, the connector 310, the first set of electrical contacts 312, and a cap 504. The first acoustic package 108a further includes a snout 510, a portion of which is positioned in the third housing portion 528.
The housing 580 also includes a third housing portion 578 having a generally cylindrical shape. Positioned within the housing 580 are various components including one or more speakers, such as a driver and a balanced armature. The housing 580 may be made of any suitable material, such as polycarbonate, and may be transparent, translucent, or opaque.
The second acoustic package 108b further includes a magnet 552, a connector 570, a set of annular electrical contacts 522 on the connector 570, and a cap 554. The second acoustic package 108b further includes a snout 560, a portion of which is positioned in the third housing portion 578. The components of the first acoustic package 108a and the second acoustic package 108b may be attached or coupled using any suitable means, such as adhesives, mechanical fasteners, ultrasonic welding, and the like.
Example dimensions of the first acoustic package 108a are as follows. In front elevational view, the housing 530 may have a length from an extremity of the first housing portion 508 to an extremity of the second housing portion 506 of about approximately 16 mm to about approximately 18 mm, such as approximately 16.4 mm, the first housing portion 508 may have a width of about approximately 12 mm to about approximately 14 mm, such as approximately 13.1 mm, and the second housing portion 506 may have a width of about approximately 6 mm to about approximately 8 mm, such as approximately 7.3 mm. In side view, the housing 530 may have a height of about approximately 8 mm to about approximately 10 mm, such as approximately 8.8 mm. The third housing portion 578 may have an outside diameter of about approximately 4 mm to about approximately 6 mm, such as approximately 5.0 mm. The second acoustic package 108b may have similar example dimensions.
Other embodiments of the acoustic package may have a different shape. For example, in one embodiment, an acoustic package may have a housing that has an asymmetric teardrop shape in front elevational view. The housing of the acoustic package for the left ear may be larger towards the left of the housing and smaller towards the right of the housing in front elevational view. Similarly, the housing of the acoustic package for the right ear may be larger towards the right of the housing and smaller towards the left of the housing in front elevational view. Other shapes for the acoustic package that fit the anatomy of an ear are possible. Accordingly, the first acoustic package 108a may have any suitable configuration and corresponding dimensions that fits a left ear and the second acoustic package 108b may have any suitable any suitable configuration and corresponding dimensions that fits a right ear.
The in-ear canal microphone 608 may be utilized as an error reference microphone for feedback active noise cancellation. The driver 606 may serve as a woofer and may provide a suitable low-frequency response. The balanced armature 610 may serve as a tweeter and may provide a suitable high-frequency response. In some embodiments the first acoustic package 108a includes only analog components, which may allow for long usage of the first acoustic package 108a.
The first housing portion 508 and the second housing portion 506 of the housing 530 define a first housing cavity 638 and a second housing cavity 636. The flexible printed circuit board 604, the driver 606, the balanced armature 610, the in-ear canal microphone 608, and the balanced armature port 612 may be positioned at least partially in the first housing cavity 638 and the second housing cavity 636. The snout 510 is generally cylindrically shaped and includes a snout proximal portion 628 positioned at least partially within the third housing portion 528 and a snout distal portion 626. A first wing 616a and a second wing 616b positioned at the snout proximal portion 628 may function to secure the snout proximal portion 628 to the third housing portion 528.
The snout 510 further includes a first flange 618, an intermediate snout portion 620, and a second flange 622. The first flange 618 is positioned flush against the third housing portion 528. The snout 510 has a first opening at the snout proximal portion 628, a second opening at the snout distal portion 626, and a snout passage therebetween such that the sound emitted by driver 606 and the balanced armature 610 may pass through the first opening, the snout passage, and the second opening.
The snout 510 further includes one or more acoustic mesh layers 624, which may be made of any suitable material, such as stainless steel, and have oleophobic and hydrophobic properties. The one or more acoustic mesh layers 624 may provide suitable acoustic resistance so that the effect of external acoustics on the components of the first acoustic package 108a is reduced or eliminated. The one or more acoustic mesh layer 624 may also not interfere with sound generated by the first acoustic package 108a and allow air to pass through.
The first ear interface 106a includes the proximal portion 340, the upper portion 342, and the distal portion 344. The first ear interface 106a further includes a first opening 642 at the proximal portion 340 and a first cavity 632 and a second cavity 630 extending away, or inwardly, from the first opening 642. The first cavity 632 has a partial generally capsule shape generally matching the partial generally capsule shape of the first housing portion 508. The second cavity 630 has a partial generally capsule shape generally matching the second partial generally capsule shape of the second housing portion 556. The matching shapes of the first cavity 632 and the second cavity 630 allow the first acoustic package 108a to be positioned within the first ear interface 106a.
The first ear interface 106a further includes a second opening (not illustrated in
The memories of the first electronics package 104a and/or the second electronics package 104b may store instructions that may be executed by the microcontroller and/or the DSPs. In some embodiments, the memory may store virtual auditory display filters that the microcontroller processor and/or the DSPs may apply to audio signals that the first electronics package 104a and the second electronics package 104b receive.
One or more components of the first electronics package 104a and/or the second electronics package 104b (for example, the IMU-based sensor system, the magnetometer, an accelerometer, a gyroscope, or other suitable components) may be configured to capture head orientation data for a head orientation of a wearer of the first ear-worn device and the second ear-worn device. The memories of the first electronics package 104a and/or the second electronics package 104b may store instructions that when executed by the microcontroller and/or the DSPs may cause the microcontroller and/or the DSPs to determine the head orientation of the wearer based on the head orientation data captured by the one or more components.
The memories of the first electronics package 104a and/or the second electronics package 104b may store instructions that when executed by the microcontroller and/or the DSPs may also cause the microcontroller and/or the DSPs to perform active noise cancellation of the sounds of the wearer captured by the in-canal microphone in the first acoustic package 108a and/or the second acoustic package 108b and/or active noise cancellation of external sounds captured by the high signal-to-noise ratio analog microphone in the first electronics package 104a and/or the second electronics package 104b.
Electronic components of the first electronics package 104a, such as the microcontroller, may control electronic components of the second electronics package 104b. Accordingly, the first electronics package 104a may be considered as a primary electronics package and the second electronics package 104b may be considered as a secondary electronics package.
The virtual auditory display device 100 may interface with a virtual auditory display system as described herein to render virtual auditory display sound for a wearer. Virtual auditory display sound may be considered as immersive, panoramic sound that is capable of being perceived as emanating from any point in space surrounding the wearer. The virtual auditory display system may apply digital filters to a multi-channel audio signal to generate audio signals that are sent to the virtual auditory display device 100. The virtual auditory display device 100 generates, based on the audio signals, the virtual auditory display sound.
Virtual auditory display sound may be especially desirable for a music creator, who may create music with multiple channels (for example, 9.1.6). The music creator may use a digital audio workstation to create music. The music creator may then use the virtual auditory display system to apply digital filters to the music so that the music may be rendered as virtual auditory display sound by the virtual auditory display device 100. The virtual auditory display system may allow the music creator to apply different acoustic environment filters so that the virtual auditory display sound may be heard in different simulated acoustic environments (for example, in a car, in a night club).
The music creator may then remix or modify the music based on their hearing the music rendered as virtual auditory display sound. For example, the music creator may move the location of certain sounds, emphasize certain sounds, deemphasize certain sounds, and the like. In this fashion, the music creator may utilize the virtual auditory display device 100 as part of an iterative process of creating music until the music creator achieves the desired effect for the music.
The virtual auditory display device 100 may switch between a transparency mode and an immersion mode by detecting interactions of the wearer with the virtual auditory display device 100. For example, the wearer may tap once on either the first electronics package 104a or the second electronics package 104b to cause the virtual auditory display device 100 to enter a transparency mode. In the transparency mode one or more microphones of the first electronics package 104a and the second electronics package 104b may capture external sounds, signals based on the captured external sounds may be generated, and the one or more speakers of the first acoustic package 108a and the second acoustic package 108b may output sound based on the generated signals.
As another example, the wearer may also tap twice on either the first electronics package 104a or the second electronics package 104b to cause the virtual auditory display device 100 to enter the immersion mode. In the immersion mode the first electronics package 104a and the second electronics package 104b may perform active noise cancellation on signals corresponding to sound captured by one or more microphones of the first electronics package 104a and the second electronics package 104b.
The virtual auditory display device 100 may provide other functionality enabled by components of the virtual auditory display device 100. For example, the arrays of microphones may enable the virtual auditory display device 100 to selectively amplify certain sounds and selectively cancel certain sounds. The DSPs may enable sound detection and the arrays of microphones may enable the virtual auditory display device 100 to locate detected sounds. The virtual auditory display device 100 may then amplify the detected sounds. As another example, the microphones of the virtual auditory display device 100 may enable active noise cancellation of sounds generated by the wearer, such as the wearer's voice. The virtual auditory display device 100 may also provide functionality other than what is described herein.
The electronics package 1104 includes a microphone cover 1218 including multiple perforations 1216, a first proximity sensor window 1220a and a second proximity sensor window 1220b. The electronics package 1104 also includes a magnet 1206 positioned inward relative to the microphone cover 1218. The microphone cover 1218 and the magnet 1206 form a generally cylindrical recess 1202 having a generally planar surface 1204. A set of electrical connectors 1208 extend outwards from the generally planar surface 1204. Each electrical connector is configured to connect with an electrical contact of the set of annular electrical contacts 1212. In some embodiments, the set of electrical connectors 1208 includes seven electrical connectors. In some embodiments, the set of electrical connectors 1208 includes a set of seven pogo pins.
The generally cylindrical recess 1202 has the same general shape as the magnet 1214 and the connector 1210. The electronics package 1104 may thus removably couple to the acoustic package 1108. The electronics package 1104 is removably magnetically coupleable to the acoustic package 1108 due to attractive magnetic forces between the magnet 1214 and the magnet 1206.
The electronics package 1104 also includes an antenna component 1330 and a battery 1332. The antenna component 1330 may include one or more antennas configured to receive and transmit wireless signals (for example, Wi-Fi, Bluetooth, cellular signals). The battery 1332 may be a pouch or coin cell battery and provide power to multiple electronics components of the electronics package 1104. The battery 1332 may be or include a rechargeable battery. In some embodiments, the electronics package 1104 may be removed from the ear-worn device and placed in a charging case having electrical contacts that connect with the set of electrical connectors 1208, so that the charging case may charge the electronics package 1104.
The electronics package 1104 also includes a printed circuit board 1308 which may include multiple electronics components such as a microcontroller, memory, codecs with audio digital signal processors (DSPs), multiple microphones, and the set of electrical connectors 1208. In some embodiments, the printed circuit board 1308 includes nine microphones. The electronics package 1104 further includes a circuit board 1334, which may include multiple electronics components such as an inertial measurement unit (IMU)-based sensor system (which may be referred to as an IMU), a magnetometer, and/or other sensors, such as accelerometers and/or gyroscopes to aid head orientations detections, and proximity sensors to detect proximity to a wearer. The electronics package 1104 further includes one or more sensor lenses 1336.
The electronics package 1104 also includes the magnet 1206 and a housing 1324 including a microphone manifold. The microphone manifold may have a continuously increasing radius like the microphone manifold 418 so as to reduce or eliminate a Helmholtz resonance. The electronics package 1104 further includes the microphone cover 1218, which includes the multiple perforations 1216. The multiple perforations 1216 allow sounds to pass through and be captured by the multiple microphones on the printed circuit board 1308. The microphone cover 1218 may be made of any suitable material, such as stainless steel, and the multiple perforations 1216 may be created by chemical etching. In some embodiments a perforation has a diameter of approximately 150 microns.
The electronics package 1104 further includes an electrical connector spacer 1312, which functions to space apart the electrical connectors of the set of electrical connectors 1208, and a glide film layer 1316, which functions to reduce a friction of the generally planar surface 1204.
Although not depicted in
The other virtual auditory display device that comprises the ear-worn device 1102 and a corresponding ear-worn device for the right ear may provide the same functionality as the virtual auditory display device 100. In addition, the other virtual auditory display device may provide acoustic zoom functionality that may be static, adaptive, or directional. For example, the other virtual auditory display device may enhance sounds based on a head orientation of the wearer. As another example of additional functionality, the other virtual auditory display device may provide gradual noise cancellation when starting or playing sound received from another device. Other functionality will be apparent.
In addition, the electronics package 1104 of the ear-worn device 1102 may be rotatable relative to the acoustic package 1108. The ear-worn device 1102 may thus provide a rotatable user interface for the wearer. The wearer may thus rotate the electronics package 1104 to cause the ear-worn device to perform certain functions, such as to adjust a sound volume. Other functionality will be apparent.
As described herein, an acoustic package (for example, the first acoustic package 108a) may be removed from the ear interface (for example, the first ear interface 106a). The ear interface may be made of a material (for example, silicone) that may be subject to degradation from use. On some occasions, the acoustic package may not be positioned within the ear interface as securely as desired. Accordingly, a mechanism for further securing the acoustic package within the ear interface may be desirable.
The collar 700 is removably coupleable to the first ear interface 106a. When coupled to the first ear interface 106a, the collar 700 extends generally circumferentially around the proximal portion 340 proximate to the first opening 642. The shelf portion 704 abuts the cap 554 and the lifter portion 710 abuts a portion of the first ear interface 106a. The inner perimeter of the collar 700 may be generally the same or slightly smaller than the outer perimeter of the proximal portion 340 proximate to the first opening 642, so as to result in the collar 700 exerting pressure upon the proximal portion 340.
A function of the collar 700 is to further secure the first acoustic package 108a in the first ear interface 106a. The first wing portion 706, the second wing portion 708a, and the third wing portion 708b may mate with corresponding notches or grooves in the proximal portion 340 and assist with the securing function. Accordingly, the collar 700 may further secure the first acoustic package 108a within the first ear interface 106a. The collar 700 may also be used with the second ear interface 106b and the other ear interfaces described herein.
The ear interfaces described herein may provide a generally sealed acoustic fit between the ear interface and the ear in which the ear interface is positioned. The generally sealed acoustic fit may prevent exterior noises from reaching the ear canal, thus preventing exterior noises from interfering with the sound produced by the acoustic package.
However, changes in air pressure, such as when the wearer of a virtual auditory display device is changing elevation (for example, when the wearer is traveling in an airplane) may affect the sound produced by the acoustic package. For example, the dynamic driver may become pinned and unable to move, thus affecting the low frequency response of the acoustic package.
The pressure-equalization vent 908 also includes a second opening 906 at the passage 1234 and a passage 902 between the first opening and the second opening 906. The diameter of the first opening or the second opening 906 may be approximately 0.1 mm. The diameter of the first opening or the second opening 906 may be approximately 0.1 mm, although the diameter of the first opening or the second opening may have other suitable sizes to allow for air passage. The passage 902 may have a diameter of approximately 0.1 mm at certain portions of the passage 902. A hollow plug 904 is positioned in the passage 902. The hollow plug 904 may function to ensure that air may travel through the passage 902. Air may pass through the first opening, the passage 902, and the second opening 906 to allow for static air pressure equalization between an air pressure in an ear canal of the wearer and an exterior air pressure without affecting the sound produced by the acoustic package. Furthermore, the pressure-equalization vent 908 may be sized to prevent undue ingress of particulate matter.
Although
The pressure-equalization vent 908 may have any passage size or combination of passage sizes that allows for air flow between the exterior and the cavities of the ear interface 1106 while maintaining the desired acoustic resistance properties.
Furthermore, in some embodiments, the pressure-equalization vent 908 may include one or more layers of acoustic mesh. For example, the pressure-equalization vent 908 may have a passage size larger than 0.1 mm and include one or more layers of acoustic mesh. It will be appreciated that the pressure-equalization vent 908 may have varying passage sizes and/or configurations to allow for the passage of air while achieving desired acoustic resistance properties.
The pressure-equalization vent 952 may also include a first layer of acoustic mesh 954a and a second layer of acoustic mesh 954b. The acoustic mesh 954a and the acoustic mesh 954b are positioned between the cap 504 and the flexible printed circuit board 604. Each layer of acoustic mesh may be generally circular and have a rayl value of 900. The two layers of acoustic mesh may function as acoustic resistors that prevent interference from external noises while still allowing air to pass through the passage 956. The layers of acoustic mesh act as acoustic resistors in parallel, so that their values are additive. Acoustic meshes that have other rayl values may be utilized in some embodiments to provide the desired acoustic resistance. The pressure-equalization vent 952 may allow for static air pressure equalization between an air pressure in the second housing portion 506 of the housing 530 and an exterior air pressure without unduly affecting the acoustic performance of the first acoustic package 108a or allowing undue transmission of unwanted acoustic energy. Furthermore, the pressure-equalization vent 952 may be sized to prevent undue ingress of particulate matter.
The configuration of the pressure-equalization vent 952 may be varied to achieve static pressure equalization across a range of operational environments for the first acoustic package 108a, while providing sufficient resistance to acoustic energy. For example, the diameter of the passage 956 may be reduced and total resistive value of the one or more layers of acoustic mesh may be reduced in order to achieve these objectives. As another example, the diameter of the passage 956 may be increased and the number of layers of acoustic mesh may be increased in order to achieve these objectives. It will be appreciated that the pressure-equalization vent 952 may have varying passage sizes and/or configurations to allow for the passage of air while achieving desired acoustic resistance properties.
Although
The graph 800 indicates a difference in level between the second audio signal 806 and the third audio signal 808 of approximately five (5) decibels (dB) at a frequency region 802 between 63 hertz (Hz) and 125 Hz. Accordingly, the third acoustic package has an improved frequency response relative to that of the second acoustic package. However, the graph 800 indicates that that each of the frequency response of the second audio signal 806 and the frequency response of the third audio signal 808 generally corresponds to the frequency response of the first audio signal 804 throughout wide ranges of frequencies.
The graph 850 indicates a difference in level between the second audio signal 856 and the first audio signal 862 of approximately 25 dB at a frequency region 852 between approximately 125 Hz and approximately 250 Hz. Accordingly, the frequency response of the second acoustic package does not track the frequency response of the first acoustic package well in the frequency region 852.
The graph 850 further indicates a difference in level between the second audio signal 856 and third audio signal 858 of approximately 10 dB at a frequency region 854 between approximately 250 Hz and approximately 500 Hz. Accordingly, the third acoustic package, with the pressure-equalization vent 952 having a single layer of acoustic mesh, has an improved frequency response relative to that of the second acoustic package with the pressure-equalization vent 952 but without any layers of acoustic mesh.
The graph 850 further indicates the frequency response of the fourth audio signal 860 generally corresponds to the frequency response of the first audio signal 862. Accordingly, the fourth acoustic package, with the pressure-equalization vent 952 having a double layer of acoustic mesh, has an improved frequency response relative to that of the second acoustic package and the third acoustic package.
Therefore, an ear-worn device having one or more pressure-equalization vents such as the pressure-equalization vent 908 and the pressure-equalization vent 952 may have desirable acoustic properties. Such an ear-worn device may be able to maintain a quality of the audio produced by the ear-worn device while blocking external sounds from being heard by a wearer and interfering with the audio.
Furthermore, an ear-worn device with one or more pressure-equalization vents such as the pressure-equalization vent 908 and the pressure-equalization vent 952 may allow for air pressure differentials between an air pressure in the cavities of the ear-worn device and an external air pressure to be reduced or eliminated. Such an ear-worn device may thus be more comfortable to wear during times when a wearer may experience changes in air pressure, such as when traveling in an airplane.
Each of the first ear-worn device 1502a and the second ear-worn device 1502b includes an electronics package (shown as a first electronics package 1504a and a second electronics package 1504b), an ear interface (shown as a first ear interface 1506a and a second ear interface 1506b), and an acoustic package (shown individually as a first acoustic package 1508a and a second acoustic package 1508b). Certain components of each of the first ear-worn device 1502a and the second ear-worn device 1502b may be generally similar to certain components of the ear-worn device 1102 of
The virtual auditory display device 1500 may be utilized for various purposes, such as providing passive noise protection for military personnel and first responders, as well as active noise enhancement. The virtual auditory display device 1500 may provide additional functionality.
Certain components of each of the first ear-worn device 1602a and the second ear-worn device 1602b may be generally similar to certain components of the first ear-worn device 102a and of the second ear-worn device 102b of
The virtual auditory display device 1600 may be utilized for various purposes, such as for providing hyper-realistic training of military personnel and facilitating multi-threaded communications amongst military personnel.
The first ear-worn device 1702a also includes an ear cup 1714a coupled to the main housing 1710a. A first electronics package 1704a is positioned in a center of the ear cup 1714a and is coupled to the ear cup 1714a by multiple bands 1734a. The first ear-worn device 1702a also includes a first ear interface 1706a that includes a first acoustic package 1708a.
The second ear-worn device 1702b includes a main housing 1710b, an ear cup 1714b coupled to the main housing 1710b, and a compute housing 1722b connected to the main housing 1710b by multiple rigid connectors 1724b. The second ear-worn device 1702b also includes two sets of microphones and a volume control knob (not illustrated in
The first electronics package 1704a and the second electronics package 1704b may be generally similar to the electronics package 1104 of the ear-worn device 1102 of
Each of the compute housing 1722a and the compute housing 1722b may include multiple electronics components (for example, one or more processors, memory). The electronics package of the first ear-worn device 1702a may be electrically connected to the multiple electronics components of the compute housing 1722a. The second electronics package 1704b may be electrically connected to the multiple electronics components of the compute housing 1722b. Compute functionality may be shared between the electronics package and the multiple electronics components in the compute housing.
A wearer may insert the first ear interface 1706a into a left ear and the second ear interface 1706b into a right ear. The wearer may then place the first ear-worn device 1702a over the wearer's left ear and the second ear-worn device 1702b over the wearer's right ear, and the attractive magnetic forces between the first acoustic package 1708a and the first electronics package 1704a and between the second acoustic package 1708b and the second electronics package 1704b will cause the first acoustic package 1708a and the first electronics package 1704a to couple together and the second acoustic package 1708b and the second electronics package 1704b to couple together.
The generally sealed acoustic fit of the first ear interface 1706a and of the second ear interface 1706b, in combination with the ear cup 1714a and the ear cup 1714b, may provide passive ear protection for a wearer of the first ear-worn device 1702a and the second ear-worn device 1702b. The microphones on the main housing 1710a and the main housing 1710b may capture external sounds and use active noise cancellation technology to provide active ear protection for the wearer.
Additionally or alternatively, the first ear-worn device 1702a and/or the second ear-worn device 1702b may amplify external sounds captured by the microphones. The virtual auditory display device 1700 may thus enhance the hearing of the wearer of the first ear-worn device 1702a and the second ear-worn device 1702b, allowing the wearer to hear sounds that the wearer may not otherwise hear.
A wearer of the virtual auditory display device 1700 may turn the controller 1800 on and off using the switch 1814 and adjust an audio volume using the volume control knob 1808. When on, the controller 1800 may control audio interactions that a wearer of the virtual auditory display device 1700 may have. For example, the virtual auditory display device 1700 may have a transparency mode and an active noise cancellation mode. The wearer may select the transparency mode by selecting a button 1806a labeled “AUI” (auditory user interface). In this mode the virtual auditory display device 1700 may enhance the sounds captured by the first set of microphones on the virtual auditory display device 1700, thereby allowing the wearer to hear sounds that the wearer may not otherwise hear.
The wearer may also select either a button 1806b labeled “1”, a button 1806c labeled “2”, or a button 1806d labeled “3”. Selecting one of these three buttons may switch the virtual auditory display device 1700 into the active noise cancellation mode and cause the virtual auditory display device 1700 to emit audio received from an external device corresponding to the selected button. The wearer may push the push-to-talk button 1804 to transmit communications to the external device.
The form factor, ergonomics, and ability to switch between an enhanced hearing mode and multiple communications channels may make the virtual auditory display device 1700 and the controller 1800 suitable for challenging use cases, such as military use cases.
Each of the first ear-worn device 2202a and the second ear-worn device 2202b includes an electronics package (shown as a first electronics package 2204a and a second electronics package 2204b), an ear interface (shown as a first ear interface 2206a and a second ear interface 2206b), and an acoustic package (shown individually as a first acoustic package 2208a and a second acoustic package 2208b). Certain components of each of the first ear-worn device 2202a and the second ear-worn device 2202b may be generally similar to certain components of the ear-worn device 1102 of
The virtual auditory display device 2200 may provide the auditory user interface controls of the controller 1800. For example, a wearer of the virtual auditory display device 2200 may tap on the first electronics package 2204a and/or the second electronics package 2204b to switch between a first, transparency, mode and a second, active noise cancellation, mode. The first electronics package 2204a and/or the second electronics package 2204b may be touch-sensitive. Accordingly, the wearer may be able to move between different communications channels using touch gestures.
The first ear interface and the second ear interface of the virtual auditory display device 1500, the virtual auditory display device 1600, the virtual auditory display device 1700, and the virtual auditory display device 2200 may be custom made for the wearer's ears so as to provide a generally acoustically sealed fit for the wearer's ears. Accordingly, the virtual auditory display device 1500, the virtual auditory display device 1600, the virtual auditory display device 1700, and the virtual auditory display device 2200 may thus provide passive protection for wearer's ears.
Each of the virtual auditory display device 1500, the virtual auditory display device 1600, the virtual auditory display device 1700, and the virtual auditory display device 2200 may include components that allow for wireless transmission and reception of signals over IP-based networks and/or radio communication networks, such as those used by military personnel and public safety officers.
One advantage of the virtual auditory display devices described herein is that the virtual auditory display devices may enable wearers to hear sounds that are beyond the visible spectrum, hear sounds that may be too high or too low for human ears, and feel vibrations that may be too subtle to notice. For example, military personnel, such as military personnel on a battlefield, in a rescue mission, or performing normal duties, may utilize the virtual auditory display devices to detect threats before they become visible, hear enemy movements before they're audible, and feel changes in the environment before they're noticeable. Such virtual auditory display devices may provide military personnel with a competitive advantage and allow them to complete their missions with greater efficiency and safety.
As an example, military personnel could use the sensory augmentation features of the devices to identify the location of an enemy target. One possible scenario is that an enemy has made a brief sound that can be identified by digital signal processing. Examples of these sounds could be a gunshot, a footstep or other noise. The devices may identify the location of the sound and inform the military personnel the location of the target. Guidance to the military personnel could include a notification or continuous sound beacon placed in virtual auditory space or a spoken notification from a conversational user interface.
Another advantage of the virtual auditory display devices described herein is that such virtual auditory display devices may allow wearers to communicate with others in even loud environments and reduce interference. The virtual auditory display devices may filter out background noise and amplify the wearer's voice, thereby allowing the wearer to easily communicate with others and reducing distractions. The virtual auditory display devices may allow wearers to communicate with others clearly and effectively even if the wearers are in loud environments such as concerts and construction sites.
An HRTF may be for one person. Generating an individual HRTF typically requires a highly specialized environment and acoustic testing equipment. A person must remain still for approximately 30 minutes in an anechoic chamber while audio signals are emitted from different known locations. A microphone is placed in each ear of the person to capture the audio signals. However, this method presents challenges as there may be spurious responses due to factors such as the chamber, the audio signal source(s) and the microphone, that need to be eliminated in order to obtain an accurate Head Related Impulse Response (HRIR) which can then be converted to an HRTF. Furthermore, any movement by the person may affect the measurements, which may result in an inaccurate HRTF for the person. Another practical limitation of measuring an HRIR is that the time to collect directly scales with the number of discrete coordinates and practically limits the resolution of the resulting HRTF.
So-called universal HRTFs have been utilized to overcome disadvantages of individual HRTFs. Such universal HRTFs may be produced by averaging or otherwise combining measurements from multiple persons. However, such combining typically results in losing the individual characteristics of each person that are necessary to produce accurate virtual 3D sound for the person. As a result, such universal HRTFs may not accurately locate sound in virtual 3D space for all users, especially sound that is located directly in front of a user at approximately zero degrees azimuth and zero degrees elevation.
Another prior approach has attempted to simulate a personalized HRTF using photogrammetry of the head, torso, and pinna, or using other methods with highly precise head, torso, and pinna scanning via time of flight or structured light. A physical acoustics model is then generated based on the resulting scanned form. However, this approach may not yield convincing rendering of virtual 3D space, because after the physical scan is measured, the physics-based simulation of sound interacting with the modeled surface may introduce complexity and inaccuracy in the resulting psychoacoustic cues.
The technology described herein provides technical solutions to the technical problems of the prior approaches described above. The technology may utilize virtual auditory display filters that may result in accurately rendered sounds in their locations in virtual auditory space. The virtual auditory display filters may utilize spectral shaping techniques, using equalizers, filters, and/or dynamic range compression, to manipulate the frequency spectrum of audio signals. Virtual auditory display filters may be generated without resort to direct physical measurements (for example, measurements in an anechoic chamber, photogrammetry, etc.).
Virtual auditory display filters may be or include functions that manipulate a frequency spectrum of an audio signal. Virtual auditory display filters may be or include digital filters, such as parametric equalization (EQ) filters that allow for adjustment of parameters such as the center frequency, gain, quality (Q or q), cutoff frequency, slope, bandwidth and/or filter type. The parameters may be set as a function of a location of a sound in virtual auditory space. The functions or the digital filters may affect the frequency spectrum of an audio signal by creating notches and peaks in the audio signal. The notches, peaks, and other spectral shaping of the audio signal accurately places the resulting sound in virtual auditory space. Furthermore, the notches, peaks, and other spectral shaping of the audio signal produces a processed audio signal that may be used to output high-quality clear sound that, in the example of music recordings, may accurately represent the original recorded performance and allow listeners to hear subtleties and nuances of the original recorded performance. As described herein, a digital filter may refer to a digital filter, a function, and/or some combination of one or more functions or one or more digital filters.
Virtual auditory space may be described as a virtual 3D sound environment of a person in which the person may perceive a sound as emanating from any location in the virtual 3D sound environment. In the described technology, each location in virtual auditory space may have an associated function or digital filter that is applied to audio signals that have that location. The application of the function or digital filter to an audio signal with a location results in sound, which may be referred to as virtual auditory display sound, that is perceived by the person as coming from that location. Accordingly, the person, who may be wearing headphones, earbuds, or other ear-worn devices, may experience virtual auditory display sound. Other advantages of the described technology will be apparent.
The virtual auditory display system 2302 may include a binauralizer 2338. The binauralizer may include a system memory 2318, which may include a left ear digital filter map 2320a and a right ear digital filter map 2320b. The binauralizer 2338 may also include a left ear convolution engine 2316a, a right ear convolution engine 2316b, and a spatialization engine 2314. The virtual auditory display system 2302 may also include other components, modules and/or engines, such as those described with reference to, for example,
In some embodiments, the virtual auditory display system 2302 may be or include a software application that may execute on a digital device. A digital device is any device with at least one processor and memory. Digital devices are discussed further herein, for example, with reference to
The virtual auditory display device 2300 may include a first ear-worn device 2302a and a second ear-worn device 2302b. The first ear-worn device 2302a and the second ear-worn device 2302b may each be any ear-worn, ear-mounted or ear-proximate device such as an earphone of a pair of earphones, an earbud of a pair of earbuds, a headphone of a headset, a speaker of a virtual reality headset, and the like. In some embodiments, the virtual auditory display device 2300 may be an embodiment of the virtual auditory display devices described herein. The first ear-worn device 2302a and/or the second ear-worn device 2302b may include components, such as an inertial measurement unit (IMU), an accelerometer, a gyroscope, and/or a magnetometer, that detect a head orientation of a wearer wearing the first ear-worn device 2302a and the second ear-worn device 2302b.
In some embodiments, a digital device (for example, a laptop or desktop computer) may receive an encoded audio file 2306 that has one or more channels of audio. Examples of an encoded audio file 2306 include 2.0 (two channels of audio), 2.1 (three channels of audio), 5.1 (six channels of audio), 7.1.4 (12 channels of audio), and 9.1.6 (16 channels of audio). The digital device may decode the encoded audio file 2306 to obtain decoded audio objects 2308 and an input audio signal 2312 that includes one or more audio sub-signals (alternately, audio channels). Each of the decoded audio objects 2308 and/or the audio sub-signals may have associated coordinates which identify the location of the audio object in virtual auditory space. The coordinates may be cartesian coordinates, spherical coordinates, and/or polar coordinates. Although specific examples of encoded audio files are described herein, the technology is not limited to such examples, and may be used with audio files that have any number of channels.
The digital device may send the coordinates 2310 to the spatialization engine 2314 and the input audio signal 2312 to the left ear convolution engine 2316a and the right ear convolution engine 2316b. In some embodiments, the virtual auditory display system 2302 receives the encoded audio file 2306 and decodes the encoded audio file 2306 to obtain the decoded audio objects 2308 and the input audio signal 2312.
As described with reference to, for example,
As described with reference to, for example,
The spatialization engine 2314 may determine, based on the acoustic environment 2332, a first acoustic environment digital filter and a second acoustic environment digital filter. An acoustic environment digital filter may be or include a digital filter that is applied to an audio signal to manipulate the audio signal so as to produce the effect of the audio being played, generated or produced in a particular acoustic environment. The spatialization engine 2314 may provide the first acoustic environment digital filter to the left ear convolution engine 2316a and the second acoustic environment digital filter to the right ear convolution engine 2316b.
While the virtual auditory display system 2302 is receiving the input audio signal 2312, one or both of the first ear-worn device 2302a and the second ear-worn device 2302b may detect a head orientation of a wearer of the virtual auditory display device 2300 and provide the head orientation and an audio source distance 2326 (which may be specified by the wearer) to the virtual auditory display system 2302.
The binauralizer 2338 may, for each audio sub-signal of the one or more audio sub-signals, obtain multiple first processed audio sub-signals and multiple second processed audio sub-signals. The binauralizer 2338 may do so by determining, based on the virtual auditory space location associated with the audio sub-signal and the head orientation, a particular first location in the virtual auditory space for the audio sub-signal. The left ear digital filter map 2320a maps locations in virtual auditory space to digital filters and/or functions for the first ear-worn device 2302a and the right ear digital filter map 2320b maps locations in virtual auditory space to digital filters and/or functions for the second ear-worn device 2302b.
Virtual auditory display filters may be or include functions and/or digital filters that the virtual auditory display system 2302 applies to audio signals to create virtual auditory display sound. A generation system, discussed in more detail with reference to, for example,
The binauralizer 2338 may select a particular first digital filter and/or function from the left ear digital filter map 2320a and a particular second digital filter and/or function from the right ear digital filter map 2320b in the system memory 2318. The binauralizer 2338 may provide the particular first digital filter and/or function to the left ear convolution engine 2316a and the particular second digital filter and/or function to the right ear convolution engine 2316b.
The left ear convolution engine 2316a may apply the particular first digital filter and/or function and the first acoustic environment digital filter to the audio sub-signal to obtain a first processed audio sub-signal. The left ear convolution engine 2316a may then generate, based on the multiple first processed audio sub-signals, an output audio signal 2322a for the first ear-worn device 2302a. The right ear convolution engine 2316b may apply the particular second digital filter and/or function and the second acoustic environment digital filter to the audio sub-signal to obtain a second processed audio sub-signal. The right ear convolution engine 2316b may then generate, based on the multiple second processed audio sub-signals, an output audio signal 2322b for the second ear-worn device 2302b. The graph 2324a depicts an example impulse response for the output audio signal 2322a and the graph 2324b depicts an example impulse response for the output audio signal 2322b.
The communication module 2402 may send requests and/or data between components of the virtual auditory display system 2302 and any other components or devices, such as the virtual auditory display device 2300 and a generation system 2580 (described with reference to, for example,
The audio input module 2404 may receive the input audio signal 2312 from, for example, the general purpose computing device on which the virtual auditory display system 2302 executes. The audio output module 2406 may provide the output audio signal 2322a to the first ear-worn device 2302a and the output audio signal 2322b to the second ear-worn device 2302b.
The calibration and personalization module 2408 may calibrate IMUs and/or other sensors of the first ear-worn device 2302a and the second ear-worn device 2302b. The calibration and personalization module 2408 may also generate personalization audio signals and receive personalization information for personalizing filters. The user interface module 2410 may provide user interfaces that allow users to, among other things, select an acoustic environment, select an audio visualization, control audio volume, and request calibration and/or personalization procedures be performed by the virtual auditory display system 2302.
The data storage 2420 may include data stored, accessed, and/or modified by any of the engines, components, modules or the like of the virtual auditory display system 2302. The data storage 2420 may include any number of data storage structures such as tables, databases, lists, and/or the like. The data storage 2420 may include data that is stored in memory (for example, random access memory (RAM)), on disk, or some combination of in-memory and on-disk.
The memory 2450 may store software and/or firmware. The IMU sensor system 2452 and/or the magnetometer 2454 may detect a head orientation of a wearer of the virtual auditory display device 2300 and/or user interactions with the virtual auditory display device 2300. The microcontroller 2456 may execute software and/or firmware stored in the memory 2450 or in the storage of the microcontroller 2456.
The power management component 2458 may provide power management. The audio DSP 2460 may process audio signals to perform functions such as noise cancellation. The microphones 2462 may capture audio, such as environmental audio and/or audio from a wearer of the first ear-worn device 2302a. The speakers 2464 may output sound based on the output audio signal 2322a and the output audio signal 2322b.
The first ear-worn device 2302a and/or the second ear-worn device 2302b may include components other than those depicted in
An engine, component, module, or the like of the virtual auditory display system 2302, the first ear-worn device 2302a, the second ear-worn device 2302b, or a generation system 2580 (described with reference to, for example
Although a limited number of engines, components, and modules are depicted in
A digital filter may be or include one or more parametric equalization (EQ) filters that allow for adjustment of parameters such as the center frequency, gain, quality (Q or q), cutoff frequency, slope, bandwidth and/or filter type. The parametric EQ filters may be or include biquad filters. The biquad filters may be or include peaking, low shelf, and high shelf filters. In some embodiments, the digital filters may be or include one or more finite impulse response (FIR) filters. The FIR filters may be generated from or based on one or more infinite impulse response (IIR) filters. In some embodiments, the digital filters may be or include one or more IIR filters, or any other suitable type of digital filter.
The digital filters that the generation system 2580 generates may be organized into multiple groups. The groups of digital filters may include a group of notch filters, a group of head shadow filters, a group of shelf filters, a group of peak filters, a group of beam filters, a group of stereo filters, a group of rear filters, a group of top filters, a group of top transition filters, a group of bottom filters, a group of bottom transition filters, and a group of broadside filters. Other groups are possible. Certain digital filters or groups of digital filters may be utilized for purposes of setting the locations of sounds in virtual auditory space (for example groups of notch filters). Certain digital filters or groups of digital filters may be utilized for purposes of ensuring that sounds meet required thresholds of tonal quality, clarity, brightness, and the like.
A digital filter may be or include an algorithm and one or more parameters for the algorithm. For example, the algorithm may be or include a high shelf, a low shelf, and a peaking algorithm. The one or more parameters may be or include a center frequency, a quality (Q or q), a gain, and a sampling frequency. For example, a notch digital filter may specify a peaking algorithm, an initial center frequency of 6600 Hz, a Q of 15, and an initial gain of −85 decibels (dB). The one or more parameters may be modified. For example, an initial center frequency may be shifted to obtain a shifted center frequency and an initial gain may be modified by a parameter modifier (see, for example, the discussion with reference to
Digital filters may be generated and utilized based on how the digital filters represent individuals' geometries interact with sound waves. For example, a digital filter having a high shelf algorithm may produce a high shelf that may be a virtual representation of how the geometry of individual's concha bowl interacts with sound waves.
The method 2500 may include a step 2502 of calibration and/or personalization. The virtual auditory display system 2302 (for example, the calibration and personalization module 2408) may perform calibration of the IMUs and/or other sensors of the virtual auditory display device 2300 using various devices and/or services 2526, such as one or both of the first ear-worn device 2302a and the second ear-worn device 2302b, a cloud-based computing service, and/or a peripheral to a computing device, such as a camera.
The virtual auditory display system 2302 (for example, the calibration and personalization module 2408) may perform personalization of the virtual auditory display system using various methodologies and/or techniques, such as: 1) a user-directed action and/or perception of acoustic cues 2516; 2) acoustic quality user feedback 2518; 3) anatomical measurements 2520; 4) demographic information 2522; and 5) audiometric measurements 2524.
User-directed action and/or perception of acoustic cues 2516 may include capturing responses of a user to locations of acoustic cues. Responses may be vocal responses of a user captured using a microphone of a computing device, gestures (for example, head and/or arm movements) of a user captured using a camera of a computing device and/or one or both of the first ear-worn device 2302a and the second ear-worn device 2302b, and user input captured via a graphical user interface (GUI) of a computing device.
Acoustic quality user feedback 2518 may include user-directed feedback on acoustic quality (for example, responses to questions on quality metrics such as brightness, warmth, clarity, etc., responses to questions provided by a GUI or an Audio User Interface (AUI)), and observations of user behavior such as user song and/or notification preferences via, for example, a GUI or AUI.
Anatomical measurements 2520 may include measurements of user anatomical features, such as the head, the pinna, and/or the concha, via scanning or prediction. Anatomical measurements 2520 of one or more users may also include direct measurements (for example, via silicone ear impressions) and indirect measurements obtained via sensors and/or computer peripherals.
Demographic information 2522 may include information provided by users such as user age or other demographics and a digital fingerprint of a user generated from one or more user features such as age, gender, and/or other user characteristics.
Audiometric measurements 2524 may include those provided by user input and or obtained via acoustic measurements, such as in an anechoic chamber while audio signals are emitted from different known locations.
The method 2500 may include a step 2504 of the generation system 2580 (for example, a model generation module 2586 of the generation system 2580, see
For each model of the multiple models 2570, for each location in virtual auditory space, the generation system 2580 may generate one or more first digital filters (for the left ear) and one or more second digital filters (for the right ear) based on the model. Accordingly, for the multiple models 2570, for each location in virtual auditory space, the generation system 2580 may generate multiple first digital filters and multiple second digital filters.
For example, for the one or more head and torso models 2560, the generation system 2580 may generate one or more first digital filters and one or more second digital filters that take into account shoulder width and/or breadth, head diameter, neck height, and other factors. For the one or more concha models 2558 the generation system 2580 may generate one or more first digital filters and one or more second digital filters that represent the acoustic effects of the physical features of the concha. These features include (but are not limited to), concha depth, width, and angle.
As another example, for the one or more pinna models 2556, the generation system 2580 may generate one or more first digital filters and one or more second digital filters that represent the acoustic effects of the physical features of the pinna. These features include (but are not limited to), pinna height, width, depth, location on the head, and flare angle relative to head. For the one or more canal models 2562, the generation system 2580 may generate one or more first digital filters and one or more second digital filters that take into account the physical proportions of the pinna, concha, and other ear components.
Also at the step 2504, for each location in virtual auditory space, the generation system 2580 may sum, aggregate, or otherwise combine the multiple first digital filters into combined first digital filters, and may sum, aggregate, or otherwise combine the multiple second digital filters into combined second digital filters. The combined first digital filters may be or include one or more finite impulse response (FIR) filters. The combined second digital filters may also be or include one or more FIR filters. Accordingly, at the conclusion of the step 2504, for all the locations in virtual auditory space, there may be a set of combined first digital filters and a set of combined second digital filters.
At a step 2506, the generation system 2580 may generate a mapping or association of the combined first digital filters to their corresponding locations in virtual auditory space for the left ear. The generation system 2580 may also generate a mapping or association of the combined second digital filters to their corresponding locations in virtual auditory space for the right ear. The generation system 2580 may utilize cartesian, polar, and/or spherical polar coordinates for the mapping or association.
At a step 2508, the generation system 2580 may generate a file, a database, or other data structure that includes the mapping or association of the combined first digital filters to their corresponding locations in virtual auditory space and the mapping or association of the combined second digital filters to their corresponding locations in virtual auditory space.
At a step 2510, the generation system 2580 may provide or store the file, the database, or other data structure on one or more non-transitory computer-readable media of a device. The device may be the first ear-worn device 2302a and/or the second ear-worn device 2302b, a mobile device such as a phone or a tablet, a laptop or desktop computer, another device, or any combination of the foregoing.
At a step 2512, the virtual auditory display system 2302 may select the combined first digital filters and the combined second digital filters for use. After selection, at a step 2514, the virtual auditory display system 2302 may utilize the combined first digital filters and the combined second digital filters in various applications, such as to render music. Various applications of the disclosed technology are discussed with reference to, for example,
In some embodiments, at step 2504, for each model of the multiple models 2570, the generation system 2580 may generate one or more first digital filters and one or more second digital filters for each azimuth and elevation combination at locations in virtual auditory space of one degree increments of azimuth and elevation at a distance of one meter (1 m) from a center point representing a virtual listener in virtual auditory space. The one degree increments of azimuth are from approximately negative 180 degrees, inclusive, to approximately positive 180 degrees, inclusive. The one degree increments of elevation are from approximately negative 90 degrees, inclusive, to approximately 90 degrees, inclusive. Accordingly, there are 65,160 combinations of azimuth and elevation, and therefore 65,160 locations in virtual auditory space, each location being at a distance of 1 m from the center point. Therefore, the generation system 2580 may generate 65,160 sets of one or more first digital filters and 65,160 sets of one or more second digital filters.
In some embodiments, the method 2500 may include a step of the generation system 2580 reducing the number of locations in virtual auditory space for which digital filters are generated or stored. For example, after step 2504, the generation system 2580 may a select a proper subset from the set of combined first digital filters and a proper subset from the set of combined second digital filters.
In embodiments where there are 65,160 locations in virtual auditory space, the generation system 2580 may select a proper subset from the set of combined first digital filters that includes approximately 7,000, such as 7,220, combined first digital filters. Similarly, the generation system 2580 may select a proper subset from the set of combined second digital filters that includes approximately 7,000, such as 7,220, combined second digital filters.
The generation system 2580 may select a proper subset that adequately represent locations in virtual auditory space, while reducing the amount of storage required for the sets of digital filters and reducing the amount of time to select and process digital filters. The generation system 2580 may achieve these objectives in other ways, such as by generating mapping or associations for a reduced number of locations in virtual auditory space or storing the mapping or associations for a reduced number of locations in virtual auditory space.
In some embodiments, at step 2504 the generation system 2580 does not sum, aggregate, or otherwise combine the multiple first digital filters into combined first digital filters and the multiple second digital filters into combined second digital filters. Accordingly, at the conclusion of the step 2504, for all the locations in virtual auditory space, there may be a set of multiple first digital filters and a set of multiple second digital filters. A proper subset of the set of multiple first digital filters and a proper subset of the set of multiple second digital filters may be utilized as described herein.
In such embodiments, at step 2506 the generation system 2580 may instead generate a mapping or association of the multiple first digital filters to their corresponding locations in virtual auditory space for the left ear and generate a mapping or association of the multiple second digital filters to their corresponding locations in virtual auditory space for the right ear.
Further in such embodiments, at step 2508 the generation system 2580 may instead generate a file, a database, or other data structure that includes the mapping or association of the multiple first digital filters to their corresponding locations in virtual auditory space and the mapping or association of the multiple combined second digital filters to their corresponding locations in virtual auditory space.
In some embodiments, the generation system 2580 generates multiple sets of digital filters for the locations in virtual auditory space. The generation system 2580 may generate a first set of digital filters for the left ear and a first set of digital filters for the right ear as described herein. The generation system 2580 may then generate one or more second sets of digital filters for the left ear and one or more second sets of digital filters for the right ear based on the first set of digital filters for the left ear and the first set of digital filters for the right ear. Each pair of sets may be for a different archetype representing a different user population or grouping of users.
The generation system 2580 may generate the one or more second sets of digital filters for the left ear and the one or more second sets of digital filters for the right ear by modifying one or more parameters of the digital filters for the left ear and the digital filters for the right ear. For example, the generation system 2580 may modify the center frequency of notch filters that are included in the first set of digital filters for the left ear and the first set of digital filters for the right ear. The generation system 2580 may modify the center frequency of notch filters to personalize digital filters to a user, as described with reference to, for example,
In some embodiments, the generation system 2580 may generate a first set of digital filters for the left ear and a first set of digital filters for the right ear for a distance of 1 m from a center point representing a virtual listener in virtual auditory space, as described herein. The generation system 2580 may generate one or more second sets of digital filters for the left ear and the one or more second sets of digital filters for the right ear for other distances from the center point. The generation system 2580 may generate one or more second sets of digital filters for the left ear based on the first set of digital filters for the left ear and one or more second sets of digital filters for the right ear based on the first set of digital filters for the right ear. For example, the generation system 2580 may increase the gain of digital filters for distances closer than 1 m from the center point and may decrease the gain of digital filters for distances further than 1 m from the center point. Other methods will be apparent.
The communication module 2582 may send requests and/or data between components of the generation system 2580 and any other systems, components or devices, such as the virtual auditory display system 2302. The communication module 2582 may also receive requests and/or data between components of the generation system 2580 and any other systems, components or devices.
The filter generation module 2584 may generate digital filters and the acoustic environment digital filters. A filter may be or include one or more algorithms and, optionally, one or more parameters for the one or more algorithms.
The model generation module 2586 may generate, modify, or access multiple models. The parameter generation module 2588 may generate parameters for digital filters.
The parameter mask module 2590 may generate parameter modifier masks. The parameter mask module 2590 may use image processing techniques to generate parameter modifier masks. The parameter mask module 2590 may determine one or more parameter modifiers to one or more parameter of filters using the parameter modifier masks. The parameter mask module 2590 may modify the one or more parameter using the one or more parameter modifiers.
The digital filter tuning module 2592 may receive parameters for digital filters from users and modify digital filters based on the received parameters. The user interface module 2594 may provide user interfaces that allow users to, among other things, listen to sound output from audio signals generated by application of digital filters and modify parameters of digital filters.
The data storage 2596 may include data stored, accessed, and/or modified by any of the engines, components, modules or the like of the generation system 2580. The data storage 2596 may include any number of data storage structures such as tables, databases, lists, and/or the like. The data storage 2596 may include data that is stored in memory (for example, random access memory (RAM)), on disk, or some combination of in-memory and on-disk.
The first audio signal is for a first sound that has a first location in virtual auditory space at a distance of one (1) meter (m), zero degrees (0°) azimuth and zero degrees (0°) elevation. The second audio signal is for a second sound that has a second location in virtual auditory space at a distance of one (1) m, five degrees (5°) azimuth and zero degrees (0°) elevation The third audio signal is for a third sound that has a third location in virtual auditory space at a distance of one (1) m, ten degrees (10°) azimuth and zero degrees (0°) elevation.
In the examples depicted in
For example,
Returning to
That is, based on the location (as specified by, for example, azimuth and elevation) of the sounds in virtual auditory space, the virtual auditory display system 2302 may determine the center frequencies for one or more notches in a frequency spectrum of the audio signals that cause the first ear-worn device 2302a and the second ear-worn device 2302b to produce the sounds. The virtual auditory display system 2302 may determine the center frequencies of the first notches in the frequency spectrums of the audio signals by accessing the distribution 2700 and the distribution 2750. The virtual auditory display system 2302 may determine the center frequencies of the second notches and subsequent notches in the frequency spectrums of the audio signals based on the distribution 2700 and the distribution 2750 and on one or more shifts from the center frequencies obtained from the distribution 2700 and the distribution 2750.
In some embodiments, in addition to or as an alternative to utilizing the distribution 2700 and/or the distribution 2750, the virtual auditory display system 2302 may utilize one or more center frequency curves, each of which may be for a different azimuth value, like the center frequency curve 2802 of
The parameter modifier masks depicted in
The parameter mask module 2590 may use image processing algorithms to create continuous transitions of values between the particular region and the other regions to generate the parameter modifier masks with the parameter modifier values. For example, the parameter mask module 2590 may use image processing algorithms such as a gaussian function, a sharpening function, a contrast adjustment function, a color correction function, a thresholding function, an edge detection function, and/or a segmentation function. In some embodiments, the parameter mask module 2590 uses a gaussian blur mask to generate the parameter modifier values. The parameter mask module 2590 may generate the parameter modifier mask for a right ear and then reflect the parameter modifier mask for the right ear about a vertical axis at an azimuth value of zero (0) to obtain the parameter modifier mask for the right ear.
The filter generation module 2584 may utilize the parameter modifier masks depicted in
Panel 3050 of
For the elevation trials, the closer the elevation delta is to 0°, the more accurate the representation of the virtual sound object. Similarly, for the azimuth trials, the closer the elevation delta is to 0°, the more accurate the representation of the virtual sound object. The user experience data shows that the digital filters improve the elevation delta from a mean of approximately 30.19° with a standard deviation of approximately 12.54° to a mean of approximately −0.03° with a standard deviation of approximately 4.12°. The user experience data also shows the digital filters improves the azimuth delta from a mean of approximately −0.64° with a standard deviation of approximately 7.76° to a mean of approximately −0.02° with a standard deviation of approximately 2.04°. The data shows that the digital filters according to some embodiments improve the accuracy and precision of virtual sound objects.
At a step 3104 the generation system 2580 (for example, the parameter mask module 2590) generates parameter modifier masks for the right ear and parameter modifier masks for the left ear (see, for example,
At a step 3106, for each location of multiple locations in virtual auditory space, the generation system 2580 (for example, the parameter generation module 2588) may generate one or more first parameters for one or more first digital filters and one or more second parameters for one or more second digital filters. The one or more first parameters may include one or more first q's, one or more first gains, and one or more first center frequencies. The one or more second parameters may include one or more second q's, one or more second gains, and one or more second center frequencies.
The generation system 2580 may utilize the parameter modifier masks for the right ear and the parameter modifier masks for the left ear to select, based on the location in virtual auditory space, one or more parameter modifiers that the generation system 2580 may use to modify one or more parameter to obtain one or more modified parameter. In some embodiments, the generation system 2580 multiplies the one or more parameter by the one or more parameter modifiers to obtain the one or more modified parameter.
The generation system 2580 may utilize the generally sigmoidal distribution of center frequencies for the right ear and/or the generally sigmoidal distribution of center frequencies for the left ear to determine one or more center frequencies for one or more notches in the frequency spectrums of the audio signal for the right ear and the audio signal for the left ear. The generation system 2580 may determine the center frequencies for the one or more notches based on the location in virtual auditory space.
At a step 3108, for each location, the generation system 2580 (for example, the filter generation module 2584) may generate one or more first digital filters including one or more first notch filters including the one or more first parameters. The generation system 2580 may utilize the one or more first q's, the one or more first gains, and the one or more first center frequencies to generate the one or more first notch filters. The one or more first notch filters are configured to produce one or more first notches in a first frequency spectrum of a first audio signal according to the one or more first q's, the one or more first gains, and the one or more first center frequencies when the generation system 2580 applies the one or more notch filters to an audio signal for the right ear.
At a step 3110, for each location, the generation system 2580 (for example, the filter generation module 2584) may generate, based on the one or more first digital filters, one or more combined first digital filters for the location. In some embodiments, the one or more first digital filters are IIR filters, and the one or more combined first digital filters are FIR filters.
At a step 3112, for each location, the generation system 2580 may store the one or more combined first digital filters in association with the location (in, for example, the data storage 2420).
At a step 3114, for each location, the generation system 2580 (for example, the filter generation module 2584) may generate one or more second digital filters including one or more second notch filters including the one or more second parameters. The generation system 2580 may utilize the one or more second q's, the one or more second gains, and the one or more second center frequencies to generate the one or more second notch filters. The one or more second notch filters are configured to produce one or more second notches in a second frequency spectrum of a second audio signal according to the one or more second q's, the one or more second gains, and the one or more second center frequencies when the generation system 2580 applies the one or more notch filters to an audio signal for the left ear.
At a step 3116, for each location, the generation system 2580 (for example, the filter generation module 2584) may generate, based on the one or more second digital filters, one or more combined second digital filters for the location. In some embodiments, the one or more second digital filters are IIR filters, and the one or more combined second digital filters are FIR filters.
At a step 3118, for each location, the generation system 2580 may store the one or more combined second digital filters in association with the location (in, for example, the data storage 2420).
At a step 3120 the generation system 2580 tests to see if there are more locations for which the generation system 2580 is to generate digital filters. If so, the method 3100 returns to step 3106. The generation system 2580 may perform the method 3100 multiple times to generate multiple sets of combined first digital filters and multiple sets of combined second digital filters. Each pair of sets of digital filters may be for a different archetype
In some embodiments, the generation system 2580 may perform the method 3100 several times to generate multiple sets of combined first digital filters and combined second digital filters. Each pair of sets may be for a different archetype representing a different user population or grouping of users.
The method 3150 begins at a step 3152, where the generation system 2580 (for example, the parameter generation module 2588) may generate a first generally sigmoidal distribution of center frequencies and a second generally sigmoidal distribution of center frequencies. At a step 3154 the generation system 2580 (for example, the parameter mask module 2590) may generate first parameter modifier masks and second parameter modifier masks.
At a step 3156, for each of multiple virtual auditory space locations, the generation system 2580 (for example, the parameter generation module 2588) may generate one or more first parameters for one or more first digital filters and one or more second parameters for one or more second digital filters. The one or more first parameters may include one or more first q's, one or more first gains, and one or more first center frequencies. The one or more second parameters may include one or more second q's, one or more second gains, and one or more second center frequencies.
At a step 3158, for each virtual auditory space location, the generation system 2580 may generate one or more first digital filters including one or more first notch filters including the one or more first parameters. Step 3158 is generally similar to step 3108 of the method 3100.
At a step 3160, for each virtual auditory space location, the generation system 2580 may store the one or more first digital filters in association with the virtual auditory space location. Step 3160 is generally similar to step 3112 of the method 3100.
At a step 3162, for each virtual auditory space location, the generation system 2580 may generate one or more second digital filters including one or more second notch filters including the one or more second parameters. Step 3162 is generally similar to step 3114 of the method 3100.
At a step 3164, for each virtual auditory space location, the generation system 2580 may store the one or more second digital filters in association with the virtual auditory space location. Step 3164 is generally similar to step 3118 of the method 3100.
At a step 3166 the generation system 2580 tests to see if there are more virtual auditory space locations for which the generation system 2580 is to generate digital filters. If so, the method 3100 returns to step 3156. The generation system 2580 may perform the method 3150 multiple times to generate multiple sets of one or more first digital filters and multiple sets of one or more second digital filters.
The method 3100 and the method 3150 may include additional steps. For example, the generation system 2580 may provide for testing digital filters. The generation system 2580 (for example, the user interface module 2594) may provide a user interface that allows for sound generated by audio signals to which digital filters have been applied to be played. A user may listen to the sounds and determine that one or more parameters of the digital filters should be modified. For example, the user may modify parameters of digital filters to ensure that sounds meet required thresholds of tonal quality, clarity, brightness, and the like. The generation system 2580 may provide another user interface that allows the user to modify the one or more parameters of the digital filters. The generation system 2580 (for example, the digital filter tuning module 2592) may receive the one or more parameters of the digital filters from users and modify digital filters based on the received one or more parameters.
While receiving the input audio signal, the virtual auditory display system 2302 performs step 3206 through step 3220 of the method 3200. At step 3206 one or both of the first ear-worn device 2302a and the second ear-worn device 2302b detects a head orientation of the user wearing the first ear-worn device 2302a and the second ear-worn device 2302b. The first ear-worn device 2302a and/or the second ear-worn device 2302b provide the head orientation to the virtual auditory display system 2302.
At a step 3208, for each audio sub-signal of the one or more audio sub-signals, the virtual auditory display system 2302 determines, based on the location of the audio sub-signal and the head orientation, a particular location in the virtual auditory space. At a step 3210, for each audio sub-signal, the virtual auditory display system 2302 selects, based on the particular location, particular one or more combined first digital filters and particular one or more combined second digital filters.
At a step 3212, for each audio sub-signal, the virtual auditory display system 2302 applies the particular one or more combined first digital filters to the audio sub-signal to obtain a first processed audio sub-signal. At a step 3214, for each audio sub-signal, the virtual auditory display system 2302 applies the particular one or more combined second digital filters to the audio sub-signal to obtain a second processed audio sub-signal.
At a step 3216 the virtual auditory display system 2302 tests to see if there are more audio sub-signals to process. If so, the method 3200 returns to step 3208. If not the method 3200 continues to step 3218. After processing all the audio sub-signals the virtual auditory display system 2302 obtains multiple first processed audio sub-signals and multiple second processed audio sub-signals.
At a step 3218 the virtual auditory display system 2302 generates, based on the multiple first processed audio sub-signals, a first output audio signal for the left ear-worn device, and based on the multiple second processed audio sub-signals, a second output audio signal for the right ear-worn device. The virtual auditory display system 2302 provides the first output audio signal to the first ear-worn device 2302a and the second output audio signal to the second ear-worn device 2302b.
At a step 3220 the first ear-worn device 2302a outputs first sound based on the first output audio signal and the second ear-worn device 2302b outputs second sound based on the second output audio signal. The virtual auditory display system 2302 may thus utilize the method 3200 to provide virtual auditory display sound based on an audio signal that may have multiple audio sub-signals (or channels) that would typically require multiple speakers to produce a surround sound effect. The virtual auditory display system 2302 may provide the virtual auditory display sound to the user using only the first ear-worn device 2302a and the second ear-worn device 2302b.
The method 3250 begins at a step 3252, where the virtual auditory display system 2302 (for example, the binauralizer 2338) receives a set of one or more first digital filters and a set of one or more second digital filters. At a step 3254 the virtual auditory display system 2302 (for example, the binauralizer 2338) receives an audio signal that has one or more audio sub-signals. Each audio sub-signal of the one or more audio sub-signals is associated with a virtual auditory space location.
At a step 3256 the virtual auditory display system 2302 receives a head orientation of a user. At a step 3258, for each audio sub-signal of the one or more audio sub-signals, the virtual auditory display system 2302 determines, based on the virtual auditory space location and the head orientation, a particular virtual auditory space location. At a step 3260, for each audio sub-signal, the virtual auditory display system 2302 selects, based on the virtual auditory space location or the particular virtual auditory space location, particular one or more first digital filters and particular one or more second digital filters.
At a step 3262, for each audio sub-signal, the virtual auditory display system 2302 applies the particular one or more first digital filters to the audio sub-signal to obtain a first processed audio sub-signal. At a step 3264, for each audio sub-signal, the virtual auditory display system 2302 applies the particular one or more second digital filters to the audio sub-signal to obtain a second processed audio sub-signal.
At a step 3266 the virtual auditory display system 2302 tests to see if there are more audio sub-signals to process. If so, the method 3250 returns to step 3258. If not the method 3200 continues to step 3268. After processing all the audio sub-signals the virtual auditory display system 2302 obtains multiple first processed audio sub-signals and multiple second processed audio sub-signals.
At a step 3268 the virtual auditory display system 2302 generates, based on multiple first processed audio sub-signals, a first output audio signal for a first device, and based on multiple second processed audio sub-signals, a second output audio signal for a second device. The first device may be or include, for example, the first ear-worn device 2302a, and the second device may be or include, for example, the second ear-worn device 2302b. At a step 3270 the virtual auditory display system 2302 provides the first output audio signal to the first device and the second output audio signal to the second device.
At a step 3286 the generation system 2580 (for example, the filter generation module 2584) may generate a first virtual auditory display filter and a second virtual auditory display filter. The first virtual auditory display filter may include a first set of first functions. One or more first functions, when applied to a first audio signal having a first location in virtual auditory space, may generate a first processed audio signal having a first frequency response with one or more first notches at one or more first center frequencies that are based on the first location. The one or more first notches may have one or more first peak-to-trough depths of at most −10 dB (for example, approximately-30 dB).
The second virtual auditory display filter may include a second set of second functions. One or more second functions, when applied to the first audio signal, may generate a second processed audio signal having a second frequency response with one or more second notches at one or more second center frequencies that are based on the second location. The one or more second notches may have one or more second peak-to-trough depths of at most −10 dB (for example, approximately −30 dB).
At a step 3288 the virtual auditory display system 2302 may receive an audio signal having a second location in the virtual auditory space. For example, the virtual auditory display system 2302 may receive the audio signal from a digital device on which the virtual auditory display system 2302 is executing. At a step 3290 the virtual auditory display system 2302 may receive a head orientation of a user, for example, from the virtual auditory display device 2300 that the user is utilizing.
At a step 3292 the virtual auditory display system 2302 may apply the first virtual auditory display filter, including a first subset of first functions selected based on the second location, to the second audio signal to generate a third processed audio signal having a third frequency response. At a step 3294 the virtual auditory display system 2302 may apply the second virtual auditory display filter, including a second subset of second functions selected based on the second location, to the second audio signal to generate a fourth processed audio signal having a fourth frequency response. At a step 3294 the virtual auditory display system 2302 may provide the first processed audio signal to a first sound output device (for example, the first ear-worn device 2302a) and the second processed audio signal to a second sound output device (for example, the second ear-worn device 2302b).
The virtual auditory display system 2302 may perform step 3288 through step 3294 of the method 3280 while the virtual auditory display system 2302 is receiving an input audio signal that may correspond to, for example, a song file, an audio stream, a podcast, or any other audio.
The method 3200, the method 3250, and the method 3280 may include additional steps not illustrated in
At a block 2482 the spatialization engine 2470 receives the user head orientation and audio source distance signal. At a block 2484 the spatialization engine performs an ambisonics to binaural conversion based on the output of the block 2480 and the user head orientation and audio source distance signal, by applying digital filters to the audio signals received at a block 2476 based on the locations of the audio signals in virtual auditory space. At a block 2486 the spatialization engine 2470 outputs the audio signal for the left ear and at a block 2488 the spatialization engine 2470 outputs the audio signal for the right ear.
The user interface 3300 includes an icon 3304 labeled “VAD” indicating that the virtual auditory display device 2300 is connected to the virtual auditory display system 2302 and an icon 3302 labeled “IMU” indicating that the IMU-based sensor systems of the virtual auditory display device 2300 are calibrated. The user interface 3300 also includes an encoding representation dropdown 3314 that allows the wearer to select how the virtual auditory display system 2302 should represent audio received by the virtual auditory display system 2302. Example encoding representations are mono (a single audio channel), stereo (two channels of audio), 5.1 5.1 (six channels of audio), 7.1 (eight channels of audio), 7.1.4 (12 channels of audio), and 9.1.6 (16 channels of audio).
The user interface 3300 also includes an acoustic environment dropdown 3316 that allows the wearer to select an acoustic environment in which the virtual auditory display system 2302 should render the virtual auditory display. Example acoustic environments include a dry acoustic environment, a studio acoustic environment, a car acoustic environment, a phone acoustic environment, a club acoustic environment, and a headphone acoustic environment. The virtual auditory display system 2302 may select an acoustic environment digital filter based on the selected acoustic environment and apply the acoustic environment digital filter along with the virtual auditory display filters. The virtual auditory display sounds will sound different for the wearer based on the selected acoustic environment. The user interface 3300 also includes an output volume slider 3322 allowing the wearer to adjust the volume of the sound output by the virtual auditory display device 2300.
The user interface 3300 also includes a representation 3308 of a virtual audio display. In
While outputting sounds, the virtual auditory display device 2300 detects head orientations of the wearer and sends the head orientation to the virtual auditory display system 2302. The virtual auditory display system 2302 updates the representation 3308 based on the detected head orientations. The virtual auditory display system 2302 may move the head 3312 and the sounds 3310 based on the detected head orientations.
The user interface 3300 also includes a virtual auditory display representation dropdown 3318 that allows the wearer to select how the virtual auditory display system 2302 provides the virtual auditory display representation. Example virtual auditory display representations include a custom representation (depicted in
The user interface 3300 also includes a location button 3320 which, if selected by the wearer, may cause the virtual auditory display system 2302 to change the representation 3308 such that the location specified by a certain coordinate (for example, zero degrees azimuth, zero degrees elevation) may be directly in front of the head 3312. The user interface 3300 also includes a settings icon 3306 which, if selected by the wearer, may cause the virtual auditory display system 2302 to provide an example user interface for adjusting settings for a virtual audio display.
The user interface 3350 also includes an icon 3356 labeled “VAD” indicating that the virtual auditory display device 2300 is connected to the virtual auditory display system 2302, a recommendation 3358 of a virtual auditory display filter, and a button 3360 labeled “Personalize” that the wearer may select to have the virtual auditory display system 2302 perform the personalization part of the calibration and/or personalization process. The user interface 3350 also indicates the spatialization precision estimate of the virtual auditory display of the wearer and a button 3362 labeled “Test” that the wearer may select to have the virtual auditory display system 2302 provide a test procedure for the wearer to allow the wearer to test to see if the wearer can accurately locate virtual auditory display sounds.
The user interface 3350 also includes a group 3364 of icons (labeled “A” through “G”) that indicates the set of virtual auditory display filters that create the virtual auditory display for the wearer. As depicted, the current set of virtual auditory display filters is “VAD C.” The wearer may select a different set of virtual auditory display filters by selecting a different icon in the group 3364. The wearer may then perform the calibration part of the calibration and/or personalization process by selecting the button 3354 and/or perform the personalization part of the calibration and/or personalization process by selecting the button 3360.
The user interface 3350 also allows the wearer to select a custom set of digital filters for the virtual auditory display system 2302 to use to generate the virtual auditory display. The wearer may do so by selecting a button 3368 labeled “Upload,” which allows the wearer to upload a file containing a custom set of digital filters to the virtual auditory display system 2302. The user interface element 3366 may then display the name of the file. This functionality may be desirable for wearers who already have a custom HRTF and want the virtual auditory display system 2302 to utilize the custom HRTF.
Image 3418 depicts the use of virtual auditory display filter technology in virtual monitors to provide noise isolation, sound quality, and virtualization to musicians. Image 3404 depicts the use of the virtual auditory display filter technology to mix music in any acoustic environment. Image 3420 depicts how the virtual auditory display filter technology may provide a listening experience that reinvigorates the music that listeners love for them. Image 3412 depicts using virtual auditory display filter technology in games to provide an immersive gaming experience. virtual auditory display filter technology may allow users to hear sounds emanating from locations that are not shown on users' displays and thus improve users' awareness.
Another group of example use cases relate to military, non-military (for example, first responders such as police and firefighters) and/or other organizational applications. For example, military personnel may use military radio systems to communicate with fellow soldiers, commanders, and other military personnel. The present technology may be utilized in scenarios including military operations, emergency services, aviation, marine and others.
Image 3406 depicts virtual auditory display filter technology providing improved voice pickup and voice display for organizational communications. Image 3414 depicts virtual auditory display filter technology providing augmentation of visual instrumentation with auditory signals in maritime operations. Image 3422 depicts virtual auditory display filter technology providing hyper-realistic virtual audio environments which facilitates virtual training for military and/or non-military personnel.
Image 3408 depicts virtual auditory display filter technology providing audio augmentation for orientation awareness for combat infantry situational awareness. Image 3416 depicts virtual auditory display filter technology providing audio augmentation for orientation awareness for air force orientation control. For example, pilots may use the localization of virtual beacons to assist in situational awareness. Image 3424 depicts virtual auditory display filter technology providing audio augmentation for hyper-situational awareness for unmanned aerial vehicle operations.
Another example use case of virtual auditory display filter technology involves phone calls or video conferences. For example, multiple people may talk at the same time in a phone call or video conference, making it difficult for listeners to focus on the speaker they want to hear, which may lead to confusion and misunderstandings. The present technology allows users to virtually select which talker they want to listen to through the simple movement of a head or other gesture. This attention selection mechanism may help avoid confusion and make meetings more productive.
As another example, air traffic controllers use radio messages to communicate with pilots. The air traffic controllers monitor the position, speed and altitude of aircraft in their assigned airspace visually and by radar and give directions to the pilots by radio. Often, air traffic controllers will need to communicate with multiple pilots simultaneously. Today, these situations requiring multiple pilot communications are addressed by physical switch boards that do not allow for user directed attention selection. The present technology may allow an air traffic controller to use gestures (for example, movement of a head or a hand) or other actions to localize radio communications so that there is seamless attention selection. In a simple example, multiple radio communication signals are statically arranged in unique virtual locations. The air traffic controller then looks at these predefined locations to hear the radio signal. In other examples, the radio communication signals are dynamically updated with the position, speed and altitude of the aircraft.
Other example use cases include virtual auditory display notifications to localize notifications such as voice, text-to-speech messages, email alerts, phone messages or other audio notifications, spatial navigation to use virtual auditory display cues to communicate direction and distance of virtual or real objects, which may also be used for wayfinding or orientation awareness, and spatial ambience to give a user a virtual sound environment that can be mixed with local or virtual sounds (for example, to experience music as if in a concert hall). Other example use cases of virtual auditory display filter technology are possible.
As the user 3502 moves his or her head to point towards the perceived location 3506 of the one or more sounds, one or both of the first ear-worn device 2302a and the second ear-worn device 2302b (for example, using the IMU sensor system 2452 and/or the magnetometer 2454) may detect a head orientation of the user 3502. The virtual auditory display system 2302 may utilize the detected head orientation to determine the perceived location 3506. The virtual auditory display system 2302 may then determine a delta 3508 between the location 3504 and the perceived location 3506. The virtual auditory display system 2302 may calculate the delta 3508 based on differences between the azimuth 3514 and/or elevation 3516 of the location 3504 and the azimuth 3514 and/or elevation 3516 of the perceived location 3506.
The user 3502 may user other gestures to indicate distance, such as extending an arm a specified amount, and the virtual auditory display system 2302 may use such gestures to determine the delta 3508 based on differences between the distance 3518 of the location 3504 and the distance 3518 of the perceived location 3506.
The virtual auditory display system 2302 may generate audio signals that cause the first ear-worn device 2302a and the second ear-worn device 2302b to output one or more subsequent sounds whose locations in virtual auditory space change, as indicated by solid line 3510. The user 3502 may move his or her head to follow the movement of the one or more subsequent sounds, and the perceived locations of the one or more sounds may change, as indicated by dashed line 3512.
One or both of the first ear-worn device 2302a and the second ear-worn device 2302b may detect subsequent head orientations of the user 3502 as he or she moves his or her head. The virtual auditory display system 2302 may utilize the detected subsequent head orientations to determine perceived locations of the one or more subsequent sounds. The virtual auditory display system 2302 may then determine one or more subsequent deltas between the location of the one or more subsequent sounds and the perceived locations of the one or more subsequent sounds.
The virtual auditory display system 2302 may store the deltas that the virtual auditory display system 2302 determines and utilize the deltas to modify the digital filters so as to cause the user 3502 to perceive the location of sounds in virtual auditory space to be closer to the actual locations in virtual auditory space. In some embodiments, the virtual auditory display system 2302 modifies the digital filters by selecting a different set of digital filters that the virtual auditory display system 2302 determines will reduce or minimize the deltas for the user 3502. The virtual auditory display system 2302 may then use the different set of digital filters for the user 3502.
In some embodiments, the virtual auditory display system 2302 modifies the parameters of the digital filters. For example, the virtual auditory display system 2302 may modify parameters such as center frequencies, gains, and/or q's. For example, the user 3502 may have an elevation delta of several degrees. The virtual auditory display system 2302 may modify the center frequency of a notch, a pair of notches, or a group of notches (see, for example,
The virtual auditory display system 2302 may repeat the personalization procedure one or more times until the virtual auditory display system 2302 determines that the deltas are within certain ranges or thresholds.
In some embodiments, the virtual auditory display system 2302 and/or other devices may capture other actions that the user may make in response to the user hearing a sound in virtual auditory space to indicate where the user perceives the location of the sound to be. Example actions may include vocal responses by the user, gestures by the user using parts of the user's body other than the user's head (for example, pointing with a finger or an arm of the user, waving, clapping, tapping and hand signals). Such actions may be captured by a device connected to the virtual auditory display system 2302, such as a microphone, camera or a motion sensing device.
Other example actions include the user indicating the perceived location of the sound using a graphical user interface and/or user input devices of a digital device such as a phone, tablet or laptop or desktop computer. For example, the virtual auditory display system 2302 may provide a graphical user interface that graphically represents virtual auditory space for the user, and the user may utilize an input device (mouse, keyboard, touchscreen, and/or voice command) to indicate the perceived location of the sound on the graphical representation of the virtual auditory space. It will be appreciated that there are various methods to capture user actions in response to the user perception of the location of the sound, and that the virtual auditory display system 2302, optionally in cooperation with other devices, may utilize the various methods.
At a step 3606 the virtual auditory display system 2302 selects, based on the first particular first location, particular one or more combined first digital filters from a first set of combined first digital filters and particular one or more combined second digital filters from a first set of combined second digital filters.
At a step 3608 the virtual auditory display system 2302 applies the particular one or more combined first digital filters to the personalization audio signal to obtain a first processed personalization audio signal and the particular one or more combined second digital filters to the personalization audio signal to obtain a second processed personalization audio signal.
At a step 3610 the virtual auditory display system 2302 generates, based on the first processed personalization audio signal, a first output audio signal for a left ear-worn device, and based on the second processed personalization audio signal, a second output audio signal for a right ear-worn device. At a step 3612 the left ear-worn device outputs first sound based on the first output audio signal and the right ear-worn device outputs second sound based on the second output audio signal.
At a step 3614 one or both of the left ear-worn device and the right ear-worn device detects a head orientation of a user wearing the left ear-worn device and the right ear-worn device. At a step 3616 the virtual auditory display system 2302 determines, based on the head orientation, a second particular first location in the virtual auditory space.
At a step 3618 the virtual auditory display system 2302 determines a delta between the first particular first location and the second particular first location. At a step 3620 the virtual auditory display system 2302 selects, based on the delta, a second set of combined first digital filters and a second set of combined second digital filters. The virtual auditory display system 2302 may use the second set of combined first digital filters and the second set of combined second digital filters while receiving a subsequent input audio signal.
At a step 3652 the virtual auditory display system 2302 receives a set of multiple first digital filters. At a step 3654 the virtual auditory display system 2302 receives a set of multiple second digital filters. There are one or more first digital filters and one or more second digital filters generated for each of multiple virtual auditory space locations.
At a step 3656 the virtual auditory display system 2302 receives personalization information for a user. Personalization information may include user-directed action or perception of acoustic cues, acoustic quality information, user anatomical measurements, user demographic information, and/or user audiometric measurements.
At a step 3658 the virtual auditory display system 2302 modifies, based on the personalization information for the user, the set of multiple first digital filters. At a step 3660 the virtual auditory display system 2302 modifies, based on the personalization information for the user, the set of multiple second digital filters.
In some embodiments, modifying, based on the personalization information, the set of multiple first digital filters includes modifying one or more first center frequencies of the multiple first digital filters. Moreover, modifying, based on the personalization information, the set of multiple second digital filters includes modifying one or more second center frequencies of the multiple second digital filters.
In some embodiments, modifying, based on the personalization information, the set of multiple first digital filters includes selecting a different set of multiple first digital filters. Further, modifying, based on the personalization information, the set of multiple second digital filters includes selecting a different set of multiple second digital filters.
The virtual auditory display system 2302 may provide a calibration and/or personalization process that allows a wearer of a virtual auditory display device to calibrate the virtual auditory display device and/or to personalize a virtual auditory display provided by the virtual auditory display device.
The calibration and/or personalization process may include a calibration part and a personalization part. The virtual auditory display device may include an inertial measurement unit (IMU). Calibrating the virtual auditory display device may refer to calibrating the IMU. Personalizing the virtual auditory display may refer to selecting a set of virtual auditory display filters for the wearer and/or modifying an existing set of virtual auditory display filters so that the virtual auditory display provided by the virtual auditory display device is customized to the wearer. The virtual auditory display system 2302 may allow the wearer to perform both the calibration part and the personalization part of the calibration and/or personalization process, just the calibration part, or just the personalization part.
The virtual auditory display system 2302 (for example, the user interface module 2410) may provide the user interface 3700. The wearer may start a calibration and/or personalization process by selecting a button labeled “Start” (not shown in
While the virtual auditory display system 2302 performs the calibration part of the calibration and/or personalization process, the virtual auditory display system 2302 may receive detections of head orientations of the head of the wearer from the virtual auditory display device 2300 based on data obtained from the IMU-based sensor system and/or other sensors of the first ear-worn device 2302a and/or the second ear-worn device 2302b. The virtual auditory display system 2302 may use the detections of head orientations and other factors, such as a known or estimated distance from a display providing the user interface 3700, a width and height of the display, positions of the first circle 3706a, the second circle 3706b, and/or the circle 3708, and/or other data from the IMU-based sensor system to calibrate the IMU-based sensor system.
The wearer of the virtual auditory display device 2300 may start the personalization part of the calibration and/or personalization process after completing the calibration part. The virtual auditory display system 2302 may cause the virtual auditory display device 2300 to play sounds at several locations (for example, five locations). The sounds may include, for example, sounds produced by objects that appear to the wearer as moving around his or her head, such as airplanes, helicopters, birds, and other flying creatures.
While the virtual auditory display system 2302 performs the personalization part of the calibration and/or personalization process, the virtual auditory display system 2302 may receive detections of head orientations of the head of the wearer from the virtual auditory display device 2300 based on data obtained from the IMU-based sensor system and/or other sensors of the first ear-worn device 2302a and/or the second ear-worn device 2302b. The virtual auditory display system 2302 may use the detections of head orientations and the locations of the sounds generated by the virtual auditory display system 2302 to calculate one or more deltas, as described with reference to, for example,
The virtual auditory display system 2302 may categorize the spatialization precision of the virtual auditory display for the wearer based on the estimate 3774, such as “Very Good” (
Although the virtual auditory display system 2302 is described as using circles, the virtual auditory display system 2302 may utilize other visual user interface elements in the calibration and/or personalization process. Furthermore, although the virtual auditory display system 2302 is described as receiving detections of head orientations from the virtual auditory display device 2300 in the calibration and/or personalization process, the virtual auditory display system 2302 may receive detections of head orientations from other devices connected to the virtual auditory display system 2302, such as cameras, motion sensing devices, virtual reality headsets, and the like.
One advantage of the calibration and/or personalization process is that the virtual auditory display system 2302 may personalize a set of virtual auditory display filters for a wide range of individuals. The virtual auditory display system 2302 may personalize the set of virtual auditory display filters by modifying the set of virtual auditory display filters. The virtual auditory display system 2302 may have pre-configured multiple sets of virtual auditory display filters and may modify the set of virtual auditory display filters by selecting a different set of virtual auditory display filters based on the results of the calibration and/or personalization process for a user.
Additionally or alternatively, the virtual auditory display system 2302 may modify the set of virtual auditory display filters by modifying the digital filters or functions included in the set of virtual auditory display filters. For example, where the set of virtual auditory display filters includes digital filters, the virtual auditory display system 2302 may modify parameters of the digital filters, such as the center frequencies, gains, q's, algorithm type, or other parameters based on the results of the calibration and/or personalization process for a user.
Personalization of virtual auditory display filters allows a wide range of individuals to experience immersive, accurately rendered sound in a virtual auditory space. Moreover, such individuals would not have to have HRTFs generated for them using potentially difficult and/or unreliable physical measurement procedures. Such individuals could obtain a personalized set of virtual auditory display filters simply by having the virtual auditory display system 2302 perform the calibration and/or personalization process for them. The modification of virtual auditory display filters may be performed at an initial setup procedure for the person and at any subsequent point during the person's use of the virtual auditory display system 2302 and/or virtual auditory display device 2300.
One advantage of virtual auditory display filters is that sounds in far more locations in virtual auditory space may be rendered in comparison to existing technologies. For example, a 9.1.6 configuration has 16 virtual speakers and thus may be limited to accurately rendering sounds for only those 16 virtual speaker locations. Such configurations may render sounds from other locations by smearing sounds from virtual speaker locations to represent the other locations, but such artifacts may be noticeable to listeners.
In contrast, virtual auditory display filters may be able to render sound at far more locations. For example, using locations at one degree increments of azimuth and elevation results in 65,160 locations. However, the described technology may generate virtual auditory display filters at smaller increments, resulting in even more locations at which the virtual auditory display filters may render sound. Moreover, typical approaches render sound at a modeled distance of 1 m from a center point representing the listener. The described technology may generate virtual auditory display filters for any number of distances from the center point. Accordingly, the described technology may accurately render sounds at varying distances.
One advantage of the described technology is that the described technology accurately renders virtual auditory display sound in virtual auditory space, meaning that sound is perceived by a listener as coming from the location that the creator of the sound intended for the sound. Another advantage of the described technology is that the described technology may be utilized with any ear-worn device, such as headphones, headset, and earbuds. Another advantage is that the virtual auditory display sound is high-quality and clear. Another advantage is the described technology may emphasize or deemphasize sounds in certain regions or locations of virtual auditory space so as to focus a listener's attention on those certain regions or locations. Such an approach may increase the listener's hearing abilities and allow the listener to hear sounds that the listener would not otherwise hear.
Another advantage of the described technology is that any digital device with suitable storage and processing power may store and apply the virtual auditory display filters. As described herein, a general purpose computing device such as a laptop or desktop computer may store and apply the virtual auditory display filters to audio signals to generate processed audio signals. The laptop or desktop computer may then send the processed audio signals to ear-worn devices to generate sound based on the processed audio signals. Similarly, a digital device such as a phone, tablet, or a virtual reality headset may store and apply the virtual auditory display filters to audio signals to generate processed audio signals and send the processed audio signals to ear-worn devices.
Additionally or alternatively, ear-worn devices, such as the virtual auditory display device 2300 described herein, may store and apply the virtual auditory display filters. The ear-worn devices may receive an input audio signal from, for example, a digital device with which the ear-worn devices are paired such as a phone or tablet, or from a cloud-based service. The ear-worn devices may apply the stored virtual auditory display filters to the input audio signal to generate processed audio signals and output virtual auditory display sound based on the processed audio signals. Another example is that a cloud-based service may store and apply the virtual auditory display filters to generate processed audio signals and send the processed audio signals to ear-worn devices. Other advantages will be apparent.
System bus 3812 represents one or more of any of several types of bus structures, including a memory bus or memory controller, a peripheral bus, an accelerated graphics port, and a processor or local bus using any of a variety of bus architectures. By way of example, and not limitation, such architectures include Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus, Enhanced ISA (EISA) bus, Video Electronics Standards Association (VESA) local bus, and Peripheral Component Interconnect (PCI) bus.
The digital device 3800 typically includes a variety of computer system readable media, such as computer system readable storage media. Such media may be any available media that is accessible by any of the systems described herein and it includes both volatile and nonvolatile media, removable and non-removable media.
In some embodiments, the at least one processor 3802 is configured to execute executable instructions (for example, programs). In some embodiments, the at least one processor 3802 comprises circuitry or any processor capable of processing the executable instructions.
In some embodiments, RAM 3804 stores programs and/or data. In various embodiments, working data is stored within RAM 3804. The data within RAM 3804 may be cleared or ultimately transferred to storage 3810, such as prior to reset and/or powering down the digital device 3800.
In some embodiments, the digital device 3800 is coupled to a network via communication interface 3806. The digital device 3800 can communicate with one or more networks such as a local area network (LAN), a general wide area network (WAN), and/or a public network (for example, the Internet).
In some embodiments, input/output device 3808 is any device that inputs data (for example, mouse, keyboard, stylus, sensors, etc.) or outputs data (for example, speaker, display, virtual reality headset).
In some embodiments, storage 3810 can include computer system readable media in the form of non-volatile memory, such as read only memory (ROM), programmable read only memory (PROM), solid-state drives (SSD), flash memory, and/or cache memory. Storage 3810 may further include other removable/non-removable, volatile/non-volatile computer system storage media. By way of example only, storage 3810 can be provided for reading from and writing to a non-removable, non-volatile magnetic media. The storage 3810 may include a non-transitory computer-readable medium, or multiple non-transitory computer-readable media, which stores programs or applications for performing functions such as those described herein with reference to, for example,
Programs/utilities, having a set (at least one) of program modules, such as the virtual auditory display system 2302, may be stored in storage 3810 by way of example, and not limitation, as well as an operating system, one or more application programs, other program modules, and program data. Each of the operating system, one or more application programs, other program modules, and program data or some combination thereof, may include an implementation of a networking environment. Program modules generally carry out the functions and/or methodologies of embodiments of the invention as described herein.
It should be understood that although not shown, other hardware and/or software components could be used in conjunction with the digital device 3800. Examples include, but are not limited to microcode, device drivers, redundant processing units, and external disk drive arrays, RAID systems, tape drives, and data archival storage systems, etc.
Exemplary embodiments are described herein in detail with reference to the accompanying drawings. However, the present disclosure can be implemented in various manners, and thus should not be construed to be limited to the embodiments disclosed herein. On the contrary, those embodiments are provided for the thorough and complete understanding of the present disclosure, and completely conveying the scope of the present disclosure.
It will be appreciated that aspects of one or more embodiments may be embodied as a system, method, or computer program product. Accordingly, aspects may take the form of an entirely hardware embodiment, an entirely software embodiment (including firmware, resident software, micro-code, etc.) or an embodiment combining software and hardware aspects that may all generally be referred to herein as a “circuit,” “module” or “system.” Furthermore, aspects may take the form of a computer program product embodied in one or more computer readable medium(s) having computer readable program code embodied thereon.
Any combination of one or more computer readable medium(s) may be utilized. The computer readable medium may be a computer readable signal medium or a computer readable storage medium. A computer readable storage medium may be, for example, but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any suitable combination of the foregoing. More specific examples (a non-exhaustive list) of the computer readable storage medium would include the following: an electrical connection having one or more wires, a portable computer diskette, a hard disk, a random access memory (RAM), a read-only memory (ROM), an erasable programmable read-only memory (EPROM or Flash memory), a solid state drive (SSD), an optical fiber, a portable compact disc read-only memory (CD-ROM), an optical storage device, a magnetic storage device, or any suitable combination of the foregoing. In the context of this document, a computer readable storage medium may be any tangible medium that can contain or store a program or data for use by or in connection with an instruction execution system, apparatus, or device.
A transitory computer readable signal medium may include a propagated data signal with computer readable program code embodied therein, for example, in baseband or as part of a carrier wave. Such a propagated signal may take any of a variety of forms, including, but not limited to, electro-magnetic, optical, or any suitable combination thereof.
Program code embodied on a computer readable medium may be transmitted using any appropriate medium, including but not limited to wireless, wireline, optical fiber cable, RF, etc., or any suitable combination of the foregoing.
Computer program code for carrying out operations for aspects of the present invention may be written in any combination of one or more programming languages, including an object-oriented programming language such as Java, Smalltalk, C++, Python, or the like and conventional procedural programming languages, such as the “C” programming language or similar programming languages. The computer program code may execute entirely on any of the systems described herein or on any combination of the systems described herein.
Aspects of the present invention are described below with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems) and computer program products according to embodiments of the invention. It will be understood that each block of the flowchart illustrations and/or block diagrams, and combinations of blocks in the flowchart illustrations and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general-purpose computer, special purpose computer, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer readable medium that can direct a computer, other programmable data processing apparatus, or other devices to function in a particular manner, such that the instructions stored in the computer readable medium produce an article of manufacture including instructions which implement the function/act specified in the flowchart and/or block diagram block or blocks.
The computer program instructions may also be loaded onto a computer, other programmable data processing apparatus, or other devices to cause a series of operational steps to be performed on the computer, other programmable apparatus or other devices to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide processes for implementing the functions/acts specified in the flowchart and/or block diagram block or blocks.
While specific examples are described above for illustrative purposes, various equivalent modifications are possible. For example, while processes or blocks are presented in a given order, alternative implementations may perform routines having steps, or employ systems having blocks, in a different order, and some processes or blocks may be deleted, moved, added, subdivided, combined, and/or modified to provide alternative or sub-combinations. Each of these processes or blocks may be implemented in a variety of different ways. Also, while processes or blocks are at times shown as being performed in series, these processes or blocks may instead be performed or implemented concurrently or in parallel or may be performed at different times. Further any specific numbers noted herein are only examples: alternative implementations may employ differing values or ranges.
Throughout this specification, plural instances may implement components, operations, or structures described as a single instance. Structures and functionality presented as separate components in example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein. Furthermore, any specific numbers noted herein are only examples: alternative implementations may employ differing values or ranges.
Components may be described or illustrated as contained within or connected with other components. Such descriptions or illustrations are examples only, and other configurations may achieve the same or similar functionality. Components may be described or illustrated as “coupled,” “couplable,” “operably coupled,” “communicably coupled” and the like to other components. Such description or illustration should be understood as indicating that such components may cooperate or interact with each other, and may be in direct or indirect physical, electrical, or communicative contact with each other.
Components may be described or illustrated as “configured to,” “adapted to,” “operative to,” “configurable to,” “adaptable to,” “operable to” and the like. Such description or illustration should be understood to encompass components both in an active state and in an inactive or standby state unless required otherwise by context.
The use of “or” in this disclosure is not intended to be understood as an exclusive “or.” Rather, “or” is to be understood as including “and/or.” For example, the phrase “providing products or services” is intended to be understood as having several meanings: “providing products,” “providing services,” and “providing products and services.”
It may be apparent that various modifications may be made, and other embodiments may be used without departing from the broader scope of the discussion herein. For example, the virtual auditory display system 2302 may utilize a group of FIR filters for each of certain locations in virtual auditory space and a group of IIR filters for each of other certain locations in virtual auditory space. As another example, the virtual auditory display system 2302 may provide audio signals to any device capable of directing sound towards the ears of a listener. As another example, a virtual auditory display device may be any device or set of devices (such as a pair of speakers) capable of producing sound based on output audio signals generated by the virtual auditory display system 2302.
Therefore, these and other variations upon the example embodiments are intended to be covered by the disclosure herein.
This application claims priority to U.S. Provisional Patent Application No. 63/493,699, filed on Mar. 31, 2023, and entitled “SENSORY AUGMENTATION, NOISE CANCELLATION, AND VOICE AMPLIFICATION DEVICES, AND ASSOCIATED SYSTEMS AND METHODS” and to U.S. Provisional Patent Application No. 63/498,509, filed on Apr. 26, 2023, and entitled “EAR-WORN DEVICES AND ASSOCIATED SYSTEMS, METHODS, AND NON-TRANSITORY COMPUTER-READABLE MEDIA,” and is related to U.S. patent application Ser. No. 18/621,974, filed on the same day herewith, and entitled “VIRTUAL AUDITORY DISPLAY FILTERS AND ASSOCIATED SYSTEMS, METHODS, AND NON-TRANSITORY COMPUTER-READABLE MEDIA.” Each of the foregoing applications is incorporated by reference herein in its entirety.
Number | Date | Country | |
---|---|---|---|
63493699 | Mar 2023 | US | |
63498509 | Apr 2023 | US |