This disclosure relates generally to hearing assistive devices. More specifically, this disclosure relates to generating sound level control for with hearing assistive devices.
A hearing aid or deaf aid is an electroacoustic device, which is designed to amplify sound for a user, usually with the aim of making speech more intelligible, and to correct impaired hearing as measured by audiometry. In the United States, hearing aids are considered medical devices and are regulated by the Food and Drug Administration (FDA). Therefore, ordinary small audio amplifiers or other plain sound reinforcing systems cannot be sold as “hearing aids”.
Earlier devices, known as ear trumpets or ear horns, were passive funnel-like amplification cones designed to gather sound energy and direct t into the ear canal. Similar devices can include the bone anchored hearing aid, and cochlear implant. A primary issue that can minimize the effectiveness of hearing aids is called the compression effect. The compression effect takes place when the amplification needed to make quiet sounds audible, if applied to loud sounds, damages the inner ear (cochlea). Louder sounds are therefore reduced giving a smaller audible volume range and hence inherent distortion. However, hearing protection can also be provided by an overall cap to the sound pressure and impulse noise suppression, which is available in some high-end hearing aids.
The above-described background relating to hearing aids is merely intended to provide a contextual overview of hearing aid technology, and is not intended to be exhaustive. Other context regarding hearing aids may become further apparent upon review of the following detailed description.
Non-limiting and non-exhaustive embodiments of the subject disclosure are described with reference to the following figures, wherein like reference numerals refer to like parts throughout the various views unless otherwise specified.
In the following description, numerous specific details are set forth to provide a thorough understanding of various embodiments. One skilled in the relevant art will recognize, however, that the techniques described herein can be practiced without one or more of the specific details, or with other methods, components, materials, etc. In other instances, well-known structures, materials, or operations are not shown or described in detail to avoid obscuring certain aspects.
Reference throughout this specification to “one embodiment,” or “an embodiment,” means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment. Thus, the appearances of the phrase “in one embodiment,” “in one aspect,” or “in an embodiment,” in various places throughout this specification are not necessarily all referring to the same embodiment. Furthermore, the particular features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
As utilized herein, terms “component,” “system,” “interface,” and the like are intended to refer to a computer-related entity, hardware, software (e.g., in execution), and/or firmware. For example, a component can be a processor, a process running on a processor, an object, an executable, a program, a storage device, and/or a computer. By way of illustration, an application running on a server and the server can be a component. One or more components can reside within a process, and a component can be localized on one computer and/or distributed between two or more computers.
Further, these components can execute from various computer readable media having various data structures stored thereon. The components can communicate via local and/or remote processes such as in accordance with a signal having one or more data packets (e.g., data from one component interacting with another component in a local system, distributed system, and/or across a network, e.g., the Internet, a local area network, a wide area network, etc. with other systems via the signal).
As another example, a component can be an apparatus with specific functionality provided by mechanical parts operated by electric or electronic circuitry; the electric or electronic circuitry can be operated by a software application or a firmware application executed by one or more processors; the one or more processors can be internal or external to the apparatus and can execute at least a part of the software or firmware application. As yet another example, a component can be an apparatus that provides specific functionality through electronic components without mechanical parts; the electronic components can include one or more processors therein to execute software and/or firmware that confer(s), at least in part, the functionality of the electronic components. In an aspect, a component can emulate an electronic component via a virtual machine, e.g., within a cloud computing system.
The words “exemplary” and/or “demonstrative” are used herein to mean serving as an example, instance, or illustration. For the avoidance of doubt, the subject matter disclosed herein is not limited by such examples. In addition, any aspect or design described herein as “exemplary” and/or “demonstrative” is not necessarily to be construed as preferred or advantageous over other aspects or designs, nor is it meant to preclude equivalent exemplary structures and techniques known to those of ordinary skill in the art. Furthermore, to the extent that the terms “includes,” “has,” “contains,” and other similar words are used in either the detailed description or the claims, such terms are intended to be inclusive—in a manner similar to the term “comprising” as an open transition word—without precluding any additional or other elements.
As used herein, the term “infer” or “inference” refers generally to the process of reasoning about, or inferring states of, the system, environment, user, and/or intent from a set of observations as captured via events and/or data. Captured data and events can include user data, device data, environment data, data from sensors, sensor data, application data, implicit data, explicit data, etc. Inference can be employed to identify a specific context or action, or can generate a probability distribution over states of interest based on a consideration of data and events, for example.
In addition, the disclosed subject matter can be implemented as a method, apparatus, or article of manufacture using standard programming and/or engineering techniques to produce software, firmware, hardware, or any combination thereof to control a computer to implement the disclosed subject matter. The term “article of manufacture” as used herein is intended to encompass a computer program accessible from any computer-readable device, computer-readable carrier, or computer-readable media. For example, computer-readable media can include, but are not limited to, a magnetic storage device, e.g., hard disk; floppy disk; magnetic strip(s); an optical disk (e.g., compact disk (CD), a digital video disc (DVD), a Blu-ray Disc™ (BD)); a smart card; a flash memory device (e.g., card, stick, key drive); and/or a virtual device that emulates a storage device and/or any of the above computer-readable media.
As an overview of the various embodiments presented herein, to correct for the above identified deficiencies and other drawbacks of hearing aid devices, various embodiments are described herein to facilitate the reduction of feedback related to hearing aid device.
A simplified overview is provided herein to help enable a basic or general understanding of various aspects of exemplary, non-limiting embodiments that follow in the more detailed description and the accompanying drawings. This overview is not intended, however, as an extensive or exhaustive overview. Instead, the purpose of this overview is to present some concepts related to some exemplary non-limiting embodiments in simplified form as a prelude to more detailed descriptions of the various embodiments that follow in the disclosure.
Described herein are systems, methods, articles of manufacture, and other embodiments or implementations that can facilitate the use of hearing aid devices. A variety of hearing aid devices suffer from acoustic feedback. However, the embodiments of the hearing aid device presented herein provide several advantages such as a reduced acoustic feedback and user comfort.
The absolute amplitude or the strength of an acoustic signal such as speech can depend on the distance between the source and the microphone. In a hearing assistive device or a hearing aid device, a microphone can be close to a user's mouth, but the sound sources of interest can be far from the microphone. In a typical scenario, the distance between the sources (i.e.: far-end talkers) and the microphone can range from one meter to three meters (or further), during a typical conversation. When the distance between the sound source and the microphone is very far, the amplitude capture by the microphone is expected to be very small due to propagation loss of the sound energy. When the distance between the microphone and the sound source is very near, the amplitude of the speech signal captured by the microphone will be very large. In order for the sources and the user to communicate, the device can amplify a signal from the source to a level perceptible by the user. However, when it is the user's turn to talk, the signal captured by the device's microphone can be large due to the amplification. So the user can perceive his/her own voice to be very loud and will be very uncomfortable to his/her auditory system. However if the user set his/her microphones sensitivity too low, then he/she can have problem hearing the far-end talker clearly. Therefore, the signal can be too loud for the user's ear and will cause discomfort to the user's auditory system.
A simple personal amplifier system can comprise a microphone, an amplifier, and one or more speakers. Whenever the aforementioned components are present, there is a potential for feedback. Feedback occurs when the sound from the speakers makes it back into the microphone and is re-amplified and sent through the speakers again. This loop happens so quickly that it can create its own frequency, which is heard as a howling sound. The howling sound can generally be a sinusoidal wave tone, and the distance between the microphone and the speakers can determine the frequency of the howling. The distance can control how quickly the sound can loop through the system. After repeatedly being amplified, the howling sound can be so loud that it irritates or damage one's hearing. Moreover, when the amplification gain is high in a hearing aid device, the howling sound can be very common. Therefore, an efficient howling control system is desired to improve a user's hearing experience.
Existing howling control methods are mainly focused on two aspects: (1) prevention of loop feedback from the speaker to the microphone; and (2) reduction of the amplification gain to avoid the positive feedback of the system when the howling sound occurs. Only reducing the amplification gain is not efficient even if the howling detection is adaptive because when the amplification gain is recovered to the normal value, the problem remains. Thus, when no howling sound is detected, the personal amplifier system should be reset to a normally working status. However, as long as the loop back from the speaker to the microphone exists, the howling sound will occur again when the amplification gain is recovered, creating residual feedback. Therefore, howling detection alone cannot be relied upon to mitigate the howling sound, and the user can be physically required to prevent the loop back from the speaker to the microphone.
A novel technique is proposed wherein the device can adaptively adjust the user's own voice signal to the same level of the source signals. Consequently, even after amplification by the device, the loudness of the user's own voice will be the same as far-end sources. The signal captured by the microphone can be statistically analyzed. The statistically analyzed signal can then be classified into a few classes such as very small, small, medium, large, and very large. If the signal is classified as small or medium, then the signal can be likely to be signal from the source of interest, which is far from the microphone. A signal that is classified as large or very large can be likely to be a signal from the user because the user is much closer to the microphone. The system can then adaptively reduce the amplitude of the larger signals to the same level as the small or medium level signals. Thus, the user's own voice can be perceived to be of equal loudness to the far end signal and will not be too loud for his/her auditory system.
In one non-limiting embodiment, the absolute amplitude of the signals captured by a microphone of a hearing assistive device or hearing aid device can be statistically analyzed. These signals can comprise a speech signal from multiple far-end talkers, environmental noise, and/or the user's own speech signals. The absolute amplitude level of the captured signals are statistically analyzed and classified into classes. The absolute amplitude range for each class can be empirically estimated. For example, in a quiet environment, the amplitude of a very small signal can range from 100 to 1000 counts; the amplitude can range from 1000 to 5000 counts for a small signal; the amplitude of a medium signal can range from 5000 to 10000 counts; and for a large signal, the amplitude can range from 10000 to 20000 counts and for a very large signal, the amplitude can be more than 20000 counts. The count number can be based on or related to a 16 bit quantize.
As mentioned, the absolute amplitude of the microphone-captured signal can be statistically analyzed. An adaptive statistical signal processing technique can be developed. Briefly, for a 16 bit signal, the maximum amplitude can be 32768 counts. The maximum amplitude can be divided equally into 100 bins, so each bin will be 327.68 counts linearly increased to its maximum count of 32768. For example, the first bin can be sitting on 327.68 and the second bin can be sitting on 2*327.68=655.36, and so on. So if the sample has an amplitude of 491.52 (327.68+163.84) or less will be sorted into the first bin. If the sample has an amplitude larger than 491.52 but less than or equal to 819.2 (655.36+163.84), it will be sorted into this second bin, and so on and so forth.
Therefore, the probability density of each bin can be computed. The probability density for the bin for each of the four ranges in this case is compute recursively as follows. The bin that corresponds to the amplitude of the signal at time t is added a count α. The rest of the bins will be reduced by the quantity Pm*(1−α), where m is the bin number which ranges from 1 to M, where M is the total number of bins. Pm is the probability density of bin m. This process is further illustrated by the equations below:
Bin=[B1B2B3 . . . BM], Equation (1):
wherein, the largest amplitude of a block can belong to bin B13, in this case:
P
m,t
=P
m,t-1*(1−α), Equation (2):
for m=1, 2, . . . M at time instant t
P
13,t
=P
13,t-1+α, Equation (3):
where Pm is the probability density of bin m, and P13 is probability density of bin 13. Whereas the bin number m ranges from 1, 2, 3 . . . M, excluding m=13 in this case, then P13,t-1 is the probability density of bin B13 at time frame t−1. Therefore, the sum of the probability of all the bins will be equal to one:
Σm=1MPm,t=1, where m=1,2,3 . . . M, at time framet. Equation (4):
A probability table can also be formed from the bin data. The probability table can be updated continuously for efficacy, as the device perceives various signals, because the signal environment will keep changing as the user moves from one location to another or the acoustic environment changes. The time constant for updating the probability density table can be empirically determined.
The bin with the higher probability from 1000 to 10000 is taken to be the amplitude of the sources, e.g., the speech signal(s) from far-end talkers. The bin with the higher probability from 10000 up to 32768 is deemed to be the user speech. The bin with the highest probability below 1000 is likely to be noise amplitude.
A dynamic range controller can adaptively limit the captured signals absolute amplitude levels. The absolute amplitude level is the level derived from the probability density table. The bin with the highest probability in the range from 1000 to 10000 can be deemed to be the level for the far-end talkers. Therefore, the far-end signal amplitude will not be affected by the controller, whereas the near-end signal amplitude (e.g.: the user speech amplitude) is expected to be much larger than 10000 and will be limited by the controller to the same amplitude as the far-end signal. In this way, the loudness of the user's own voice will be the same as the far-end voice from the talkers, yielding a desirable end result.
In addition, in other non-limiting embodiments, an integrated personal amplifier system can comprise acoustic feedback control to significantly improve the hearing experience. As mentioned, in personal amplifier system, the acoustic feedback can occur when there is a loop back from the speaker to the microphone. Traditional feedback/howling control systems find it difficult to recover from feedback control mode to a normal working mode. In a proposed system, an interaction between the device and the user can ensure that the feedback control system is working more efficiently.
A feedback detector can be employed after the amplifier. When the feedback occurs and is detected, the feedback detector can mute the speaker by setting the amplification gain to zero to protect the hearing and provide a warning signal by a status indicator to inform the user that the system is in a feedback protection mode. The status indicator can be a light emitting diode (LED), an audible tone, etc. After the user realizes that the feedback is occurring so that the feedback protection mode is activated, the user can verify the cause of the loop back from the speaker to the microphone. For example, the user may not have sealed the earbud into the ear canal properly, or the speaker might be placed too close to the microphone. Once the user verifies the cause, the system can recover/revert to the normal working mode in response to a user interaction with the system.
In order to make the feedback detection more accurate and reliable, a filter bank can be applied to the personal amplifier system. In each channel of the filter, there can be a feedback detector working independently; and once a feedback is detected in any frequency band, the system can go into the feedback protection mode.
According to one embodiment, described herein is a method for facilitating feedback reduction in assistive hearing devices. The method can comprise sampling, a block of sound data associated with acoustic signal samples of acoustic signals can be sampled, and generating data related to an absolute amplitude of the acoustic signal samples. Furthermore, the method can extract sample data related to a sample of the acoustic signal samples, and sort the sample data into predetermined absolute amplitude ranges.
According to another embodiment, described herein is another method for feedback reduction. The method can comprise receiving signal data related to an acoustic signal, and in response to the receiving the signal data, analyzing the signal data. The signal data can also be classified into classes, and an acoustic feedback can be detected, resulting in an acoustic feedback detection. Consequently, a channel can be muted in response to the acoustic feedback detection.
According to yet another embodiment, described herein is an apparatus for facilitating feedback reduction. The apparatus can comprise a microphone that receives first acoustic signal data related to a first acoustic signal, and an amplifier that amplifies the acoustic signal. The apparatus can also comprise an acoustic feedback detector that detects acoustic feedback signal data, and a speaker that outputs a second acoustic signal
These and other embodiments or implementations are described in more detail below with reference to the drawings.
Referring now to
Referring now to
For example if the sample amplitude is sorting into one of four ranges with more than 5000 counts but less than 10000 counts, it can sort into range 3. If the signal amplitude is less than 5000 but larger than 1500, it can be place into range 2 and so on and so for. The signal sample can be further sorted into one of the 100 bins in range 3 to form a probability table.
Referring now to
Referring now to
Referring now to
Referring now to
In order to minimize or to prevent the saturation of the far-end sources, the maximum amplitude of the dynamic range controller can be empirically set to be two bins above the bin with the highest probability. In this case, instead of the amplitude of 6226, the maximum amplitude of the dynamic range controller is set to 6881 counts.
The dynamic range controller can constrain the amplitude of all the signals to be less than or equal to the amplitude of a far-end signal. Thus, any howling can be limited.
Referring now to
In order to ensure that the far-end signal is not saturated by the dynamic controller and it is also not too loud when the user speaks, the probability distribution density of Range 3 should be considered. Referring now to
Referring now to
The above description of illustrated embodiments of the subject disclosure, including what is described in the Abstract, is not intended to be exhaustive or to limit the disclosed embodiments to the precise forms disclosed. While specific embodiments and examples are described herein for illustrative purposes, various modifications are possible that are considered within the scope of such embodiments and examples, as those skilled in the relevant art can recognize.
In this regard, while the subject matter has been described herein in connection with various embodiments and corresponding FIGs, where applicable, it is to be understood that other similar embodiments can be used or modifications and additions can be made to the described embodiments for performing the same, similar, alternative, or substitute function of the disclosed subject matter without deviating therefrom. Therefore, the disclosed subject matter should not be limited to any single embodiment described herein, but rather should be construed in breadth and scope in accordance with the appended claims below.
This patent application is a divisional application of, and claims priority to each of, U.S. Non-provisional patent application Ser. No. 15/886,078 (now U.S. Pat. No. 10,667,063), filed Feb. 1, 2018, and entitled “SOUND LEVEL CONTROL FOR HEARING ASSISTIVE DEVICES”, which is a divisional of U.S. Non-provisional patent application Ser. No. 15/097,011 (now abandoned), filed on Apr. 12, 2016, and entitled “SOUND LEVEL CONTROL FOR HEARING ASSISTIVE DEVICES”, which applications each claim further priority to each of U.S. Provisional Patent Application No. 62/278,425, filed on Jan. 13, 2016, and entitled “HEARING ASSISTIVE DEVICE WITH NEAR-END SOUND LEVEL CONTROL AND HOWLING CONTROL”, and U.S. Provisional Patent Application No. 62/218,543, filed Sep. 14, 2015, and entitled “HEARING ASSISTIVE DEVICE WITH NEAR-END SOUND LEVEL CONTROL AND HOWLING CONTROL”. The entireties of the foregoing applications are hereby incorporated by reference herein.
Number | Date | Country | |
---|---|---|---|
62278425 | Jan 2016 | US | |
62218543 | Sep 2015 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15886078 | Feb 2018 | US |
Child | 16877643 | US | |
Parent | 15097011 | Apr 2016 | US |
Child | 15886078 | US |