This application is the US National Stage of International Application No. PCT/EP2007/060710, filed Oct. 9, 2007 and claims the benefit thereof. The International Application claims the benefits of German application No. 10 2006 047 985.8 filed Oct. 10, 2006, both of the applications are incorporated by reference herein in their entirety.
The invention relates to a method for actively operating a hearing aid consisting of a single hearing device or two hearing devices. The invention also relates to a corresponding hearing aid or hearing device.
When one is listening to someone or something, disturbing noise or unwanted acoustic signals are present everywhere that interfere with the other person's voice or with a wanted acoustic signal. People with a hearing impairment are especially susceptible to such noise interference. Background conversations, acoustic disturbance from digital devices (cell phones), traffic or other environmental noise can make it very difficult for a hearing-impaired person to understand the speaker they want to listen to. Reducing the noise level in an acoustic signal, combined with automatic focusing on a wanted acoustic signal component, can significantly improve the efficiency of an electronic speech processor of the type used in modern hearing aids.
Hearing aids employing digital signal processing have recently been introduced. They contain one or more microphones, A/D converters, digital signal processors, and loudspeakers. The digital signal processors usually subdivide the incoming signals into a plurality of frequency bands. Within each of these bands, signal amplification and processing can be individually matched to the requirements of a particular hearing aid wearer in order to improve the intelligibility of a particular component. Also available in connection with digital signal processing are algorithms for minimizing feedback and interference noise, although these have significant disadvantages. The disadvantageous feature of the algorithms currently employed for minimizing interference noise is, for example, the maximum improvement they can achieve in hearing-aid acoustics when speech and background noise are within the same frequency region, making them incapable of distinguishing between spoken language and background noise. (See also EP 1 017 253 A2).
This is one of the most frequently occurring problems in acoustic signal processing, namely extracting one or more acoustic signals from different overlapping acoustic signals. It is also known as the “cocktail party problem”, wherein all manner of different sounds such as music and conversations merge into an indefinable acoustic backdrop. Nevertheless, people generally do not find it difficult to hold a conversation in such a situation. It is therefore desirable for hearing aid wearers to be able to converse in just such situations in the same way as people without a hearing impairment.
EP 1 432 282 A2 discloses a digital method for adjusting a hearing program of a hearing device to an instantaneous acoustic ambient situation, and a hearing device system for this purpose. With this method, in a digital signal analysis unit of the hearing device, characteristic auditory-based features are extracted from a digital acoustic signal and analyzed by a pattern recognizer in a signal identification unit to determine an acoustic ambient situation and generate a corresponding acoustic output signal. Said acoustic output signal is fed to a transmission unit that can be manipulated by an input unit such as a remote control, it being possible for preset parameter sets of the transmission unit to be influenced by a hearing aid wearer's input unit.
In acoustic signal processing there exist spatial (e.g. directional microphone, beam forming), statistical (e.g. blind source separation), and hybrid methods which, by means of algorithms and otherwise, are able to separate out one or more sound sources from a plurality of simultaneously active sound sources. For example, by means of statistical signal processing of at least two microphone signals, blind source separation enables source signals to be separated without prior knowledge of their geometric arrangement. When applied to hearing aids, that method has advantages over conventional approaches involving a directional microphone. Using a BSS (Blind Source Separation) method of this kind it is inherently possible, with n microphones, to separate up to n sources, i.e. to generate n output signals.
Known from the relevant literature are blind source separation methods wherein sound sources are analyzed by analyzing at least two microphone signals. A method and corresponding device of this kind are known from EP 1 017 253 A2, the scope of whose disclosure is expressly to be included in the present specification. Corresponding points of linkage between the invention and EP 1 017 253 A2 are indicated mainly at the end of the present specification.
In a specific application for blind source separation in hearing aids, this requires communication between two hearing devices (analysis of at least two microphone signals (right/left)) and preferably binaural evaluation of the signals of the two hearing devices which is preferably performed wirelessly. Alternative couplings of the two hearing devices are also possible in such an application. Binaural evaluation of this kind with stereo signals being provided for a hearing aid wearer is taught in EP 1 655 998 A2, the scope of whose disclosure is likewise to be included in the present specification. Corresponding points of linkage between the invention and EP 1 655 998 A2 are indicated at the end of the present specification.
Directional microphone control in the context of blind source separation is subject to ambiguity once a plurality of competing wanted sources, e.g. speakers, are simultaneously present. While blind source separation basically allows the different sources to be separated, provided they are spatially separate, the potential benefit of a directional microphone is reduced by said ambiguity problems, although a directional microphone can be of great benefit in improving speech intelligibility specifically in such scenarios.
The hearing aid or more particularly the mathematical algorithms for blind source separation is/are basically faced with the dilemma of having to decide which of the signals produced by blind source separation can be most advantageously forwarded to the algorithm user, i.e. the hearing aid wearer. This is basically an unresolvable problem for the hearing aid because the choice of wanted acoustic source will depend directly on the hearing aid wearer's momentary intention and hence cannot be available to a selection algorithm as an input variable. The selection made by said algorithm must accordingly be based on assumptions about the listener's likely intention.
The prior art is based on the assumption that the hearing aid wearer prefers an acoustic signal from a 0° direction, i.e. from the direction in which the hearing aid wearer is looking. This is realistic insofar as, in an acoustically difficult situation, the hearing aid wearer would look at his/her current interlocutor to obtain further cues (e.g. lip movements) for increasing said interlocutor's speech intelligibility. This means that the hearing aid wearer is compelled to look at his/her interlocutor so that the directional microphone will produce increased speech intelligibility. This is annoying particularly when the hearing aid wearer wants to converse with just one person, i.e. is not involved in communicating with a plurality of speakers, and does not always wish/have to look at his/her interlocutor.
If the direction of the wanted sound is not set to 0° for the hearing aid, ambiguity can be resolved by other additional information, e.g. by giving preference to the acoustic signal arriving with an angle of incidence that is as small as possible with respect to the forward direction. However, this severely restricts the hearing aid wearer's freedom of movement. It also creates the potential problem of ‘jumping’ between different speakers, which is unintended and experienced as unpleasant by the hearing aid wearer.
Furthermore, there is to date no known technical method for making a “correct” choice of acoustic source, or more specifically one preferred by the hearing aid wearer, after source separation has taken place.
An object of the invention is to specify an improved method for operating a hearing aid, and an improved hearing aid. In particular, an object of the invention is to determine which of the electrical output signals resulting from source separation, in particular blind source separation, is fed to the hearing aid wearer. It is therefore an object of the invention to discover which signal is, with a high degree of probability, the hearing aid wearer's preferred sound source.
The object of the invention is achieved by a method for operating a hearing aid and by a hearing aid as claimed in the claims.
Instead of leaving it up to the hearing aid to decide which of the output signals resulting from source separation is selected, i.e. compelling a hearing aid wearer to behave in a particular way, the object of the invention is achieved to the effect that this decision is left up to the hearing aid wearer him/herself.
According to the invention, a method for the active operation of a hearing aid by a hearing aid wearer is provided wherein, for tracking and selectively amplifying the hearing aid wearer's preferred acoustic signal, a “hold” command is transmitted to a signal processing section of the hearing aid which instructs the signal processing section to track said acoustic signal selected by the hearing aid wearer and to take it particularly into account in an acoustic output signal of the hearing aid.
Additionally provided according to the invention is a hearing aid with an input device, said input device being invokable and/or actuatable by a hearing aid wearer in order to actively give an instruction to the hearing aid in such a way that a “hold” command can be given by the input device to an acoustic module (signal processing section) of the hearing aid which identifies the hearing aid wearer's preferred acoustic signal which can be particularly taken into account in an output sound of the hearing aid.
The selection as to which acoustic signal or more specifically which speaker is to be tracked is made in a simple and intuitive manner by the user of the hearing aid in order to avoid the essential ambiguity of source separation methods. The hearing aid wearer knows best with whom he/she currently wishes to speak, so that no mis-classification of the preferred acoustic signal by automatic processes inside the hearing aid occurs. He/she is not restricted in his/her freedom of movement, except when executing the “hold” command, and benefits from the automatic directional microphone in an acoustically difficult situation in which a directional microphone can be extremely useful to him/her in terms of speech intelligibility.
It is inventively possible, depending on the number of microphones present in the hearing aid, to select one or more hearing aid wearer preferred acoustic signals from the ambient sound and accentuate them in the hearing aid's output sound, it being possible here to adjust a volume of the preferred acoustic signal or signals in the output sound as required.
In a preferred embodiment of the invention, the signal processing section has an unmixer module that preferably operates as a blind source separation device for separating the acoustic signals within the ambient sound. The signal processing section also has a post-processor module which, in response to the “hold” command, sets up the corresponding operating mode in the hearing aid. The signal processing section can also have a pre-processor module—the electrical output signals of which are the unmixer module's electrical input signals—which standardizes and conditions electrical signals originating from microphones of the hearing aid. In respect of the pre-processor module and unmixer module, reference is made to EP 1 017 253 A2 paragraphs [0008] to [0023].
When he/she wishes to have a particular acoustic signal tracked by the hearing aid's microphones, the hearing aid wearer enters, via the input device, the “hold” command to the signal processing section which selects the corresponding signal or signals in the post-processor module and makes it/them available to a loudspeaker of a receiver of the hearing aid at least louder than other, unwanted acoustic signals.
The input device can be any device or apparatus on the hearing aid or on a remote control for the hearing aid. In its simplest embodiment, it is a control on the hearing aid and/or remote control. However, it is likewise possible for the input device to be embodied as voice-actuated control in the hearing aid or remote control. The acoustic signal that is loudest or predominant in the ambient sound when the “hold” command is executed and/or which is preferably coming from a 0° viewing direction of the hearing aid wearer is then tracked by the source separation algorithm, preferably the BSS algorithm, and made available to the wearer in accentuated form through the receiver of the hearing aid.
According to the invention, either “simple tracking” of the preferred acoustic signal is possible in which a speaker is tracked for as long as he/she is speaking, i.e. without a memory function. Preferable, however, is so-called “intelligent tracking” whereby one or more preferred acoustic signals are analyzed and temporarily stored in the hearing aid in the form of characteristic parameters (speaker recognition). This inventively makes it possible for the speaker to be tracked by the hearing aid or more specifically the signal processing section during a genuine dialog, i.e. one in which the hearing aid wearer is speaking now and then and his/her interlocutor is silent.
Once such a “hold” operating mode is set up for a single preferred acoustic signal or the “hold” operating modes are set up for a plurality of preferred acoustic signals, its/their tracking is independent of the hearing aid wearer's head movement or viewing direction.
Further tracking of the preferred acoustic signal or signals can be inventively terminated by a release command to the input device. If a plurality of preferred acoustic signals are selected by the hearing aid wearer, it is possible here to release the preferred acoustic signal that is the loudest or predominant when the release command is executed, or preferably to release the preferred acoustic signal coming from the current 0° viewing direction of the hearing aid wearer. Independently of this, it is also possible to release any hitherto preferred acoustic signal, as the latter has already been identified. If only a single preferred acoustic signal is selected, this can naturally be dispensed with.
In addition, further tracking of the preferred acoustic signal or signals can be terminated by a time criterion (timeout) if one of the speakers or the speaker has no longer been detected for a particular time period by the hearing aid's speaker recognition system. The length of time for which the hearing aid is unintentionally left in “hold” mode can be reduced by an automatically generated release command, i.e. an automatic timeout.
Additional preferred embodiments of the invention will emerge from the other dependent claims.
The invention will now be explained in greater detail on the basis of exemplary embodiments and with reference to the accompanying drawings in which:
Within the scope of the invention (
The following description also discusses “tracking” of a preferred acoustic signal by a hearing aid wearer's hearing aid. This is to be understood as a selection made by the hearing aid wearer of one or more acoustic signals that are electrically or electronically selected by the hearing aid from other acoustic signals in the ambient sound and which are reproduced in an amplified manner compared to the other acoustic signals in the ambient sound, i.e. in a manner experienced as louder for the hearing aid wearer. For tracking of the preferred acoustic signal by the hearing aid, advantageously no account is taken of the hearing aid wearer's position in space, in particular of the hearing aid's position in space, i.e. the direction in which the hearing aid wearer is looking.
In the prior art, the microphone signals 202, 212 are mainly conditioned in three stages. In a first stage, the microphone signals 202, 212 are pre-processed in a pre-processor module 310 to improve the directional characteristic, starting with standardization of the original signals (equalizing the signal strength). In a second stage, blind source separation takes place in a BSS module 320, the output signals of the pre-processor module 310 undergoing an unmixing process. The output signals of the BSS module 320 are then post-processed in a post-processor module 330 in order to generate a desired electrical output signal 332 which is used as an input signal for a receiver 400, or more specifically for a loudspeaker 400 of the hearing aid 1, and to deliver a sound generated thereby to the hearing aid wearer. As specified in EP 1 017 253 A2, steps 1 and 3, i.e. the pre-processor module 310 and post-processor module 330, are optional.
It shall be assumed in the following that there are two acoustically independent signals 102, 104, i.e. signal sources 102, 104, in the ambient sound 100, one of said acoustic signals 102 being the acoustic signal 102 preferred by the hearing aid wearer. This preferred acoustic signal 102 is to be tracked by the hearing aid 1 or more specifically the signal processing section 300 and is to be a main acoustic component of the receiver 400 so that an output sound 402 of the loudspeaker 400 mainly contains said signal (102).
The two microphones 200, 210 of the hearing aid 1 each pick up a mixture of the two acoustic signals 102, 104—indicated by the dotted arrow (representing the preferred acoustic signal 102) and by the continuous arrow (representing the non-preferred acoustic signal 104)—and deliver them either to the pre-processor module 310 or immediately to the BSS module 320 as electrical input signals. The two microphones 200, 210 can be arranged in any manner. They can be located in a single hearing device 1 of the hearing aid 1 or distributed over both hearing devices 1. It is also possible, for instance, to provide one or both microphones 200, 210 outside the hearing aid 1, e.g. on a collar or in a pin, as long as it is still possible to communicate with the hearing aid 1. This also means that the electrical input signals of the BSS module 320 do not necessarily have to originate from a single hearing device 1 of the hearing aid 1. It is, of course, possible to implement more than two microphones 200, 210 for a hearing aid 1. A hearing aid 1 consisting of two hearing devices 1 preferably has a total of four or six microphones.
The pre-processor module 310 conditions the data for the BSS module 320 which, depending on its capability, for its part forms two separate output signals from its two, in each case mixed input signals, each of said output signals representing one of the two acoustic signals 102, 104. The two separate output signals of the BSS module 320 are input signals for the post-processor module 330 in which it is then decided which of the two acoustic signals 102, 104 will be fed out to the loudspeaker 400 as an electrical output signal 332.
For this purpose the hearing aid wearer inventively issues an appropriate command to the signal processing section 300 or more specifically the post-processor module 330, a “hold” operating mode being established for the hearing aid 1 for a particular time T by an input device 10 on or in the hearing aid 1 or by a remote control (schematically shown at bottom right in
The input device 10 can be e.g. a button 10 or a switch 10 on the hearing device 1 or on the remote control of the hearing device. It is additionally possible for the input device 10 to be embodied as a speaker recognition module in the hearing device 1 or remote control or as a voice-actuated control.
If the input device 10 is actuated, the post-processor module 330 establishes the “hold” operating mode which mainly produces the preferred acoustic signal 102 as the electrical output signal 332 of the hearing aid 1 over a particular time T. In so doing, e.g. the source signal mainly coming from the hearing aid wearer's 0° viewing direction when the input device 10 is actuated can be selected from the ambient sound 100. Other angles are also possible here. It is also possible, when the input device 10 is actuated or invoked, for the predominant or loudest signal in the ambient sound 100 to be tracked in “hold” mode. It is additionally possible for the corresponding hearing aid wearer preferred acoustic signal 102 to be identified in terms of its frequency range or more specifically its respective frequency extremes, its pitch or octave range, by a particular human voice, by music, by a particular absence of interference or by timewise similar spacings of mutually similar acoustic events or by the opposite of the above. For selecting the preferred acoustic signal 102, the acoustic signal 102 coming from the hearing aid wearer's 0° viewing direction is given preference and then tracked by the algorithm of the BSS module 320 or more specifically the post-processor module 330.
Tracking of the preferred acoustic signal 102 continues until such time as the hearing aid wearer issues a release command via the input device 10 or a speaker (corresponding to the preferred acoustic signal 102) is tracked for as long as he/she is speaking. In addition, a speaker signal 102 can be temporarily stored by a speech analyzer in the form of characteristic parameters in the signal processing section 300 or the hearing aid 1 and is tracked independently of any head movement or viewing direction of the hearing aid wearer. In the latter situation, the speaker is released either by the hearing aid wearer's release command or via a timeout.
The electrical microphone signals x1(t), x2(t), x3(t) are input signals for the BSS module 320 which separates the acoustic signals s1(t), s2(t), s3(t) respectively contained in the microphone signals x1(t), x2(t), x3(t) according to acoustic sources and feeds them out as electrical output signals s′1(t), s′2(t), s′3(t) [analogously: s′4(t), . . . , s′o(t)] to the post-processor module 330.
In the following, the hearing aid wearer prefers two acoustic signals, namely s1(t) and s3(t) (this corresponds most closely to the acoustic sources s1(t) and s3(t)). Here the hearing aid wearer successively enters the corresponding “hold” command to the input device 10 to establish the “hold” operating mode (see above), the post-processor module 330 selecting the corresponding output signals s′1(t), s′3(t) of the BSS module 320 and delivering them amplified through the receiver 400 as output sound 402, s″(t)=s″1(t)+s″2(t). Identification of the acoustic sources s1(t) and s3(t) takes place as described above.
It is additionally possible to “scan” the output signals s′1(t), s′2(t), s′3(t) of the BSS module 320 by means of the input device 10 and then make an appropriate selection.
In this embodiment of the invention it is self-evidently also possible to reproduce one or three or more preferred acoustic signals in an amplified manner.
The present specification relates inter alia to a post-processor module 20 as in EP 1 017 253 A2 which can be controlled by a hearing aid wearer via an input device (the reference numerals are those given in EP 1 017 253 A2). See also in that regard paragraph [0025] in EP 1 017 253 A2. In the invention, the pre-processor module and the BSS module can moreover be of the same design as the pre-processor 16 and the unmixer 18 in EP 1 017 253 A2. See in particular paragraphs [0008] to [0024] in EP 1 017 253 A2.
The invention also links to EP 1 655 998 A2 in providing a hearing aid wearer with stereo signals for an acoustic source selected by him/her or rather enabling a hearing aid wearer to be supplied in a binaural acoustic manner, the invention (notation according to EP 1 655 998 A2) preferably being connected downstream of the output signals z1, z2 for the right(k) and left(k) respectively of a second filter device in EP 1 655 998 A2 (see
Number | Date | Country | Kind |
---|---|---|---|
10 2006 047 985 | Oct 2006 | DE | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/EP2007/060710 | 10/9/2007 | WO | 00 | 4/7/2009 |
Publishing Document | Publishing Date | Country | Kind |
---|---|---|---|
WO2008/043758 | 4/17/2008 | WO | A |
Number | Name | Date | Kind |
---|---|---|---|
20030185411 | Atlas et al. | Oct 2003 | A1 |
20070030982 | Jones et al. | Feb 2007 | A1 |
Number | Date | Country |
---|---|---|
1017253 | Jul 2000 | EP |
1432282 | Jun 2004 | EP |
1655998 | May 2006 | EP |
1670285 | Jun 2006 | EP |
9033329 | Feb 1997 | JP |
2000066698 | Mar 2000 | JP |
2003533152 | Nov 2003 | JP |
2004512493 | Apr 2004 | JP |
2005268964 | Sep 2005 | JP |
2006178333 | Jul 2006 | JP |
WO 0187011 | Nov 2001 | WO |
Number | Date | Country | |
---|---|---|---|
20100027821 A1 | Feb 2010 | US |