N/A.
This invention relates, generally, to audio signal processing and, particularly, to systems and methods for selectively enhancing speech signals to improve speech recognition by individuals and automated processes.
The art of processing of audio signals spans a wide range of technologies and efforts. Despite the plethora of signal processing advancements related to audio signals, the processing of audio signals including or created as part of oral communications and, particularly, human speech remains a substantial challenge. For example, despite substantial investments in research and resources, speech processing and, particularly, speech recognition systems are still quite limited. These limits are due, at least in part, to the complexities of human speech and a limited understanding of natural auditory and cognitive processing capabilities. For example, the ability to recover speech information, despite dramatic articulatory and acoustic assimilation and coarticulation of speech sounds, poses substantial hurdles to enhancement of speech signals and automated processing of the underlying information communicated in speech. These hurdles are further compounded when, for example, the individual receiving the speech signals has an impairment.
Reports indicate that only about 20 percent of the more than 30 million adults with hearing loss in this country currently use hearing aids, and by 2030 there could be over 40 million adults and over 2 million children with hearing loss in the United States. The National Council on Aging indicates that untreated hearing loss of any degree has significant consequences on people's social lives, emotional health, mental health, and physical well-being. Furthermore, the Better Hearing Institute estimates that earning potential for individuals with untreated hearing loss is reduced by an average of $23,000 per year, which is twice as much for individuals with hearing aids. When multiplied by the number of American workers with hearing loss, the magnitude of total annual lost income is staggering. While many factors are related to these numbers, hearing aid performance is an important variable as indicated by the finding that only about half of all users are satisfied with how their hearing aids perform in noise. Advancements in hearing aid performance have the potential to improve quality of life for more than 10 percent of the American population as well as productivity of the average hearing-impaired worker. U.S. Pat. No. 6,732,073 to Kluender et al. provides a substantial summary of some of the difficulties and impediments to speech signal processing and enhancement and is incorporated herein by reference.
For some time, it has been understood that at least two components of sensorineural hearing loss (SNHL) reduce listeners' access to speech information. The first is a loss of sensitivity, which results in an attenuation of speech. To overcome a loss of attenuation, the signal simply needs to be made louder and noise reduced. Accordingly, many hearing aids focus on using wide dynamic range compression and various processing strategies to boost the signal-to-noise ratio, such as noise reduction and directional microphones. The second component of SNHL is a loss of selectivity, which results in a blurring of spectral detail, or distortion. Unfortunately, due to this second component of SNHL, simple amplification of speech does not necessarily improve the listeners' ability to discern the information conveyed in the speech.
Due to substantial research, it is now established that listeners with SNHL often have compromised access to frequency-specific information because spectral detail is often smeared, or blurred, by broadened auditory filters. Loss of sharp tuning in auditory filters generally increases with degree of sensitivity loss and is due, in part, to a loss or absence of peripheral mechanisms responsible for suppression. It has been learned that in the non-impaired cochlea different frequency components of a signal serve to suppress one another, and two-tone suppression has been cast as an instance of lateral inhibition. Consequently, spectral peaks in the internal representation for hearing-impaired (HI) listeners, as opposed to normal-hearing (NH) listeners, are less intense relative to spectral contrast that is reduced and more susceptible to noise. Not only are spectral peaks harder to resolve in noise due to reduced amplitude differences between peaks and valleys, but their internal representation is spread out over wider frequency regions (smeared), resulting in less precise frequency analysis, blurring between frequency varying formant patterns, and ultimately in greater confusions between sounds with similar spectral shapes.
Simultaneous spectral contrast is the intensity difference between peaks and valleys in the spectral shape of different speech sounds. Spectral peaks (formants) reflecting vocal tract resonances are important acoustic features that help define the identity of many speech sounds. A number of experimental techniques confirm that the internal representation of spectral contrast for steady state speech sounds, like vowels, is reduced in HI compared to NH listeners. For example, it has been found that peaks in vowel masking patterns for HI listeners were not resolved as well as for NH listeners, and that peak frequencies in the internal representations were often shifted away from their corresponding formant frequencies.
Decreased signal-to-noise ratios in the internal spectrum also results from auditory filters broadened by SNHL. Others found a relationship between HI listeners' estimated auditory filter bandwidths in the region of the second formant (F2) and the amount of spectral contrast needed to identify vowels in noise. These findings indicate that noise effectively reduces internal spectral contrast and that deleterious effects of noise can be offset to some extent by an increase in spectral contrast. Similarly, it has been indicated that there is a general trading relationship between spectral resolution and the amount spectral contrast needed for vowel identification.
As stated, historically, the primary function of hearing aids is to make speech in regions of hearing loss comfortably audible. Unfortunately, in this effort, hearing aids can increase the blurring of detailed frequency information by reducing internal representations of spectral contrast in at least three ways: 1) high output levels; 2) positive spectral tilt; and 3) compression (decreased dynamic range).
First, it is well known that auditory filter tuning is level dependent. Even NH listeners experience decreased frequency selectivity at high levels needed to overcome sensitivity loss for HI listeners. In ears with SNHL, high presentation levels contribute to further reductions in frequency tuning and greater smearing of spectral detail already associated with the loss of nonlinear mechanisms.
Second, hearing aids typically provide high-frequency emphasis, or a positive spectral tilt, to compensate for increases in hearing loss with frequency. However, it has been indicated that positive spectral tilt for NH listeners actually reduces the internal representation of higher frequency formants and increases the need for greater spectral contrast. Thus, it has been hypothesized that this might occur because internal representations of some formants are characterized by ‘shoulders’ rather than peaks—as a spectral ‘irregularity’ on the skirt of a more intense formant. Using an auditory filter model, it has been demonstrated that increases in spectral tilt raise the probability that a formant will be represented as a shoulder rather than a peak (similar to increases in filter bandwidth), but suppression can serve to convert (enhance) some of these shoulders into peaks. It is likely that negative effects of increased spectral tilt in NH listeners are exacerbated in HI listeners with already poor auditory filter tuning and reduced/absent mechanisms for suppression.
Third, it has long been suspected that multichannel compression in hearing aids, which is designed to accommodate different dynamic ranges of audible speech with frequency, has the potential to reduce spectral contrast and flatten the spectrum, especially when there are many independent channels and/or high compression ratios. Notably, several studies have found that compression across many independent channels increases errors for consonants differing in place of articulation, which can be highly influenced by subtle changes in spectral shape. Some have not only reported a significant decrease in vowel identification with an increase in independent compression channels, but also found that identification and number of channels were each directly related to acoustic measures of spectral contrast.
Spectral contrast is not only important for detecting differences between static spectral shapes, but also for detecting changes, which are made more subtle by coarticulation in connected speech. For example, considering the case of a formant that ends with closure silence and begins again (after closure) at a slightly higher or lower frequency. For the HI listener, there would be no perceived difference in the offset and onset frequencies, as both would be processed by the same broadened auditory filter (i.e., the change in frequency across time would be blurred). Such would not be the case for the NH listener. Instead, contrastive process operating across time would serve to “repel” these spectral prominences making them more distinct. Most conventional hearing aid processing strategies are designed to increase audibility of speech information and to improve signal-to-noise ratio by manipulating relative intensities of speech and noise. Unfortunately, these processing strategies do not adequately address the challenges of listeners with mild SNHL who experience reductions in spectral contrast as a consequence to the intensity manipulations of the processing, nor the challenges of listeners with moderate to severe hearing loss who suffer from additional reductions in spectral contrast and increased distortion arising from cochlear damage and broadened auditory filters.
Like hearing aid users, spectral blurring experienced by cochlear implant (CI) listeners is attributable to impaired cochlear/neural functioning and to device processing that is necessary to accommodate the impairment. Severe amplitude compression is needed to fit the relatively large dynamic range of speech (about 50 dB, including the effects of vocal effort) into a restricted dynamic range of electrical stimulation (often, 5-15 dB). Furthermore, a limited number of useable electrodes (typically, between 6 and 22) are available to CI listeners, who most often cannot take full advantage of even this limited spectral information provided by their electrode arrays. This is demonstrated by speech tests in quiet and in noise and by tests measuring discrimination of spectral ripples where performance as a function of number of active electrodes asymptoted at 4-7, even though the CI listeners could use a greater number in isolation for simple pitch and level discriminations. Thus, the effective number of channels for spectrally rich sounds like speech is less than the number of active electrodes.
Limited use of available spectral detail in patterns of stimulation from the CI processor is likely due to the reduced specificity of stimulation attributable to current spread, and to decreased survival and function of spiral ganglion cells. Consequently, compared to NH listeners, CI listeners need, for example, at least 4-6 dB greater spectral contrast for vowel identification in quiet and need even greater signal-to-noise ratios (SNRs) for speech in noise. Tests using NH listeners with simulated CI processing (vocoded speech) indicate that while as few as 8-12 channels might be sufficient for very good speech understanding in quiet. As many as 20 might be needed to adequately understand speech in contexts known to be exceptionally challenging for CI listeners, particularly, competing background noise, multiple talkers, and low linguistic redundancy. As with hearing aid users, transient burst onsets and rapid formant frequency changes that distinguish consonants differing in place of articulation are most troublesome for CI listeners.
To aid speech understanding in noise, some devices include noise reduction schemes and directional microphones. CI coding strategies, like spectral peak coding strategy (SPEAK) for example, analyze incoming speech into a bank of filters (e.g., 20) and use the outputs from a small number of them (e.g., 6) to stimulate corresponding places on the electrode array. CI listeners largely rely on relative differences in across-channel amplitudes to detect formant frequency information, and this is especially problematic when there is competing noise or a small number of effective channels. Furthermore, because nonlinear processes are abolished either by the impairment itself or by placement of the electrode array, natural spectral enhancement is also lost.
Thus, systems and methods for speech processing and recognition and systems and methods for manipulating audio signals including speech to improve the understanding of HI and CI listeners must balance a wide variety of variables and unknowns and continue to have long-standing need for improvement.
The present invention provides a system and method for audio signal enhancement for speech processing and/or recognition and enhancement. Unlike traditional systems, the present invention recognizes that, although counterintuitive, contrast enhancement, when applied across the entire spectrum and/or when not applied in a highly-selective or judicious manner, can actually impede a listener's or other recipient's ability to understand the underlying speech. The present invention provides a system and method to selectively manipulate or augment portions of an audio signal, for example, to allow portions of the audio signal to be enhanced and other portions of the audio signal to be unenhanced or enhanced differently. Accordingly, the present invention can be used so as to, at least, not reduce an ability of a receiving entity to process the unenhanced or differently-enhanced portions of the audio signal.
In accordance with one aspect of the present invention, a hearing aid system is provided that is configured to be coupled with an ear of an individual to selectively enhance an acoustic signal to be received by the ear of the individual. The system includes a microphone configured to receive the acoustic signal and generate an analog electrical signal responsive thereto and an analog-to-digital converter configured to receive the analog electrical signal and convert the analog electrical signal into a digital input signal. The system also includes a signal processor configured to receive the digital input signal and programmed to divide the digital input signal into a plurality of spectral channels having associated unenhanced signals. The signal processor is also configured to perform enhancement processing on a first subset of the spectral channels having associated unenhanced signals corresponding to a pathological response range of the ear of the individual and not perform enhancement processing on a second subset of the spectral channels having associated unenhanced signals outside the pathological response range of the ear of the individual. Furthermore, the signal processor is configured to combine the plurality of enhanced signals associated with each of the first subset of the spectral channels and the unenhanced signals associated with each of the second subset of the spectral channels to form a selectively enhanced output signal. The system also includes an output device configured to receive the selectively enhanced output signal and communicate the selectively enhanced output signal to the individual through the ear of the individual.
In accordance with another aspect of the present invention, a method is provided to divide an input auditory signal into a plurality of spectral channels having associated unenhanced signals and perform enhancement processing on a first subset of the spectral channels and not perform enhancement processing on a second subset of the spectral channels. The enhancement processing is performed by determining an output gain for at least the first subset of spectral channels based on a time-varying history of energy of the unenhanced signals associated with each channel in the first subset of the spectral channels and applying the output gain for each of the first subset of the spectral channels to the unenhanced signals to form enhanced signals associated with each of the first subset of the spectral channels. The system and method are then designed to combine the plurality of enhanced signals associated with each of the first subset of the spectral channels and the unenhanced signals associated with each of the second subset of the spectral channels to form a selectively enhanced output auditory signal.
In accordance with another aspect of the present invention, a system for selectively enhancing an acoustic signal is provided that includes a microphone configured to receive an acoustic signal and generate an analog electrical signal responsive thereto and an analog-to-digital converter configured to receive the analog electrical signal and convert the analog electrical signal into a digital input signal. The system also includes a signal processor configured to receive the digital input signal and programmed to divide the digital input signal into a plurality of spectral channels having associated unenhanced signals and perform enhancement processing on a first subset of the spectral channels and not perform enhancement processing on a second subset of the spectral channels. The signal processor is also programmed to combine the plurality of enhanced signals associated with each of the first subset of the spectral channels and the unenhanced signals associated with each of the second subset of the spectral channels to form a selectively enhanced output signal. The system also includes an output device configured to receive the selectively enhanced output signal and communicate the selectively enhanced output signal.
Additional features and advantages of the present invention will be apparent from the following detailed description taken in conjunction with the accompanying drawings.
The present invention provides a system and method for using contrast enhancement (CE) algorithm that is specifically designed to confine enhancement to portions of the spectrum and allow those portions to be selected and highly customized. For example, a CE algorithm may be employed that is designed to enhance spectral differences between adjacent sounds and thereby improve speech intelligibility for hearing impaired (HI) listeners by enhancing signature kinematic properties of connected speech, but is restricted to being applied to portions of the audio spectrum. The CE algorithm may be designed to achieve enhancement of spectral contrast across time, or successive spectral contrast, in addition to enhancement of simultaneous spectral contrast.
The present invention may be employed in electronic hearing aid devices for use by the hearing impaired, particularly for purposes of enhancing the spectrum such that impaired biological signal processing in the auditory brainstem is restored. This process enhances spectral differences between sounds in a fashion mimicking that of non-pathological human auditory systems. The process imitates neural processes of adaptation, suppression, adaptation of suppression, and descending inhibitory pathways, and does not impede functions that are more akin to natural, non-impaired processes by selectively controlling the enhancements. The present invention makes sounds, particularly speech sounds, more distinguishable to listeners and other receivers. Thus, the present invention is applicable to uses other than hearing aids, such as speech recognition systems.
The present invention recognizes that, for many HI listeners, amplification is used to make a signal audible, but because of limited dynamic range, spectral resolution deteriorates at amplified presentation levels. The invention addresses this problem by the manipulation of the spectral composition of the signal to overcome some of the loss of spectral resolution, and to substitute to some extent for additional amplification (which becomes deleterious at higher levels). By selectively applying such enhancements, the present invention avoids the common problems caused by enhancements applied to the entire dynamic spectrum.
Referring to
As will be described, the present invention provides a contrast enhancement algorithm and selective control mechanism designed to manipulate the spectral composition of speech sounds across time such that spectral prominences (formants) are spread apart in frequency in an effort to make them sufficiently distinct to overcome spectral blurring that occurs with a combination of SNHL, background noise, increased presentation levels, high-frequency gain, and multichannel compression. However, unlike traditional systems, the present invention recognizes that, although counterintuitive, contrast enhancement, when applied across the entire spectrum and/or when not applied in a highly selective, judicious manner, can actually impede a listener's or other recipient's ability to understand the underlying speech. To provide a high degree of contrast without a corresponding degradation or distortion created by applying contrast enhancement to, for example, portions of the spectrum that may not substantially benefit from enhancement or, when considering the entire spectrum, may ultimately reduce the overall contrast, the present invention is designed to selectively apply enhancement.
Where auditory filtering is relatively normal, any signal manipulation including contrast enhancement distorts information and perceived “naturalness.” Traditional attempts to improve speech recognition in HI listeners via simultaneous spectral enhancement employed enhancement uniformly across the spectrum, which is one likely reason for their less-than-favorable outcomes. The present invention provides systems and method for customized enhancement so that it is present, for example, only where there is significant hearing loss. For example, for listeners with mild low-frequency hearing loss sloping to moderately severe in the high frequencies, a uniform degree of enhancement might be too great in the low frequencies, thereby unacceptably distorting the signal (e.g., increasing F1 intensity too much, contributing to upward spread of masking of F2), but still insufficient in higher frequencies where it is needed most. Customization of spectral enhancement represents a significant innovation over prior methods.
Referring now to
Referring now to
Thereafter, at process block 66, channel selection for enhancement is applied. Specifically, after the input acoustic signal, x(t), is divided into a plurality of spectral channels at process block 64, channel selection for enhancement is applied such that only some of the channels are selectively enhanced. It is contemplated that this may be achieved, for example, using a block Toeplitz submatrix. The block Toeplitz submatrix may be constructed such that the spectral channels that remain unprocessed are instantiated by an identity submatrix. The channels that are selectively processed correspond to negative off-diagonal entries, for example, as illustrated in the following exemplary submatrix:
Thereafter, at process block 68, a weighted time history (e.g., 30-300 ms buffer) of the energy passing through each channel to be enhanced is converted into an RMS value. This adaptation stage can be implemented using dynamic compression with a nonlinear convex loss function, such that more recent energy passing through a channel is given greater weight than earlier occurring energy (i.e., a leaky temporal integrator).
At process block 70, the RMS value of the weighted history is converted to a gain factor for the associated channel. For example, the RMS value of the weighted history may be subtracted from unity (1) to yield a gain factor for that channel. The greater the weighted history of energy, the smaller the gain is. Maximum gain (1) is assigned when the weighted history is zero. In this way, processes of adaptation are mimicked and contribute to competition between channels.
Thereafter, at process block 72, processes of lateral inhibition are simulated. This may be achieved in the way gain is balanced across weighted frequency neighborhoods of channels. To this end, it is contemplated that a winner-take-all circuit may be used to simulate a biological network of inhibitory sidebands. Energy in a channel with a relatively high gain factor is increased at the expense of a decrease in adjacent channels with relatively low gain factors. In essence, the channel activities “compete” on a moment-by-moment basis.
The collective effects of the windowed RMS calculation (dynamic compressive gain) and lateral interactions within frequency neighborhoods results in a form of forward energy suppression specifically designed to enhance the spectrum across time. When an individual channel has relatively high energy in the past, it will tend to suppress its current energy under the condition that its neighboring channels were low in energy. This form of suppression will have the effect of sharpening dynamic modes in the spectrum, especially onsets, while flattening those that are relatively steady state, and in this way, will serve to enhance temporal contrasts. Enhancement of temporal contrasts in speech can especially aid stop consonant perception by emphasizing low-intensity transient energy characteristic of burst onsets and rapid formant transitions.
Consider the case of a single formant traversing frequency. As the formant increases in frequency, the CE algorithm successively attenuates lower-frequency filters through which the spectral prominence has already passed. This has two consequences. First, the shoulder on the low-frequency side of the formant will be sharpened because that is where the most energy was immediately prior. This will serve to “sharpen” the spectrum as compensation for “blurring” caused by an impaired cochlea. Second, the effective frequency (center of gravity) of the formant peak will be skewed away from where the formant had been before. Consequently, successive contrast will be imposed on the signal (spreading successive formants apart in frequency). It also is the case that a formant transition will be “accelerated” via this process. Because the CE algorithm successively attenuates the low-frequency shoulder, the effective slope of the processed formant becomes steeper.
The analysis and synthesis components of the above-described contrast enhancement method and circuit may employ a polyphase decomposition and oversampled discrete Fourier transformed (DFT) modulated filters. That is, as described, the input signal may be first decomposed into a plurality of subbands and CE performed within neighborhoods of subbands, then the subband process can be reversed to reconstruct the output signal. A subband scheme can utilize an analysis filter bank that splits the input into a set of M narrowband signals that are typically downsampled (decimated) by some factor N leading to more efficient processing. Intermediate processing can be performed and the constituents subsequently combined using a synthesis filter bank that is then upsampled (interpolated) by a factor of N. If no intermediate processing is performed, it is generally acknowledged that the input can be perfectly reconstructed at the output of the circuit along with some measure of pure delay. The M subband filters are derived by frequency shifting a well-constructed prototype low-pass filter h[t]. Polyphase decomposition groups the analyzing prototype filter h[t] into M subsequences prior to Fourier transformation. This segmented representation allows rearrangement of the filtering computations and increases the speed of processing approximately M-fold. The output signal is then reconstructed using a synthesis bank containing the inverse DFT matrix and the reconstruction matrix.
Referring again to
Specifically, referring to
The above-described systems and method for selective contrast enhancement may be coupled with a variety of additional processing techniques. For example, nonlinear frequency compression remaps high-frequency information above a certain start frequency into a smaller bandwidth, while leaving low frequencies below the start frequency unaltered. This represents an advance in hearing aid processing. One limitation to this new technology is that spectral contrast between peaks in the spectrum is reduced, thereby exacerbating the already limited spectral resolution of the impaired cochlea. Pre- or post-processing, frequency-compressed speech coupled with the above-described selective CE systems and methods help overcome some of this reduction in spectral contrast and allows one to effectively select the areas of compression and areas of remapped high-frequency information without disturbing areas of the spectrum that an impaired individual is capable of processing substantially normally.
Similarly, because many sources of background noise tend to be stationary and because the present CE algorithm attenuates static spectral features, noise reduction is a natural byproduct of the processing that could augment or replace existing noise reduction strategies (e.g., spectral subtraction). Along a similar line of reasoning, a persistent spectral peak associated with acoustic feedback in hearing aids could be eliminated with the CE algorithm and replace other, less desirable, feedback cancellation strategies, such as, notch filtering and a reduction in much needed high-frequency gain. Yet again, the above-described selective CE systems and methods allow one to select areas of processing and others to remain substantially unprocessed.
Thus, the present invention recognizes that impairments rarely extends across the entire frequency range of hearing. Rather, most commonly, hearing loss is most severe at specific frequencies, such as higher frequencies; although, listeners can have selective losses at other frequencies. Similarly, the present invention recognizes that normal receivers rarely benefit from enhancements or the like being applied across the full listening spectrum. For example, such “enhancement” signal processing often introduces distortion. With this recognition in place, the present invention provides a system and method to restrict contrast enhancement to only, for example, “pathological” channels or other designated channels that can benefit from enhancement without being overridden by distortion or other negative effects.
It is understood that the present invention is not limited to the specific applications and embodiments illustrated and described herein, but embraces such modified forms thereof as come within the scope of the following claims.
This invention was made with government support under Grant No. DC004072 and DC010601 awarded by the National Institute of Health. The government has certain rights in this invention.
Number | Name | Date | Kind |
---|---|---|---|
3180936 | Schroeder | Apr 1965 | A |
4051331 | Strong et al. | Sep 1977 | A |
4185168 | Graupe et al. | Jan 1980 | A |
4249042 | Orban | Feb 1981 | A |
4366349 | Adelman | Dec 1982 | A |
4396806 | Anderson | Aug 1983 | A |
4454609 | Kates | Jun 1984 | A |
4630304 | Borth et al. | Dec 1986 | A |
4630305 | Borth et al. | Dec 1986 | A |
4700361 | Todd et al. | Oct 1987 | A |
4701953 | White | Oct 1987 | A |
4852175 | Kates | Jul 1989 | A |
5027410 | Williamson et al. | Jun 1991 | A |
5029217 | Chabries et al. | Jul 1991 | A |
5388185 | Terry et al. | Feb 1995 | A |
5479560 | Mekata | Dec 1995 | A |
5742689 | Tucker et al. | Apr 1998 | A |
5793703 | Shippey | Aug 1998 | A |
6732073 | Kluender et al. | May 2004 | B1 |
8233651 | Haller | Jul 2012 | B1 |
20080144869 | Paludan-Mueller et al. | Jun 2008 | A1 |
Number | Date | Country |
---|---|---|
0556992 | Aug 1993 | EP |
Entry |
---|
Partial International Search Report, PCT/US2011/051443; 2 pages. |
Holt, L., et al., Neighboring Spectral Content Influences Vowel Identification, J. Acoust. Soc. Am. 108(2):710-722, 2000. |
Lotto, A., et al., Perceptual Compensation for Coarticulation by Japanese Quail (Coturnix coturnix japonica), J. Acoust. Soc. Am. 102(2), Pt. 1, pp. 1134-1140, 1997. |
Lotto, A., et al., General Contrast Effects in Speech Perception: Effect of Preceding Liquid on Stop Consonant Identification, Perception & Psychophysics 60(4):602-619, 1998. |
Nilsson, M., et al., Development of the Hearing in Noise Test for the Measurement of Speech Reception Thresholds in Quiet and in Noise, J. Acoust. Soc. Am. 95(2):1085-1099, 1994. |
Summerfield, Q., Auditory Enhancement of Changes in Spectral Amplitude, J. Acoust. Soc. Am. 81(3):700-708, 1987. |
Vaidyanathan, P., Quadrature Mirror Filter Banks, M-Band Extensions and Perfect-Reconstruction Techniques, IEEE ASSP Magazine, Jul. 1987, pp. 4-20. |
White, M., Compression Systems for Hearing Aids and Cochlear Prostheses, Journal of Rehabilitation Research and Development 23(1):25-39, 1986. |
Number | Date | Country | |
---|---|---|---|
20120134522 A1 | May 2012 | US |