Below, the invention is described in more detail by means of examples and the included drawings. The figures show schematically:
The reference symbols used in the figures and their meaning are summarized in the list of reference symbols. Generally, alike or alike-functioning parts are given the same or similar reference symbols. The described embodiments are meant as examples and shall not confine the invention.
The input unit 10, e.g., a microphone, receives acoustic sound 6 from the environment and outputs audio signals S1. The audio signals S1 are fed to the transmission unit 20 (e.g., a digital signal processor), which implements (embodies) a transfer function G. The audio signals are processed (amplified, filtered and so on) according to the transfer function G, thus generating output audio signals 7, which are fed to the output unit 80, which may be a loudspeaker. The output unit 80 outputs signals 8 to be perceived by a user of the hearing system 1, which may be acoustic sound (or other signals) derived from the incoming acoustic sound 6.
The audio signals S1 are also fed to the characterizing unit 40, which derives a set C1 of sound-characterizing data therefrom. This set C1 is fed to the evaluating unit 50, and the evaluating unit 50 also receives directional information D1, provided by the storage unit 60.
The evaluating unit 50 derives, in dependence of the set C1 of sound-characterizing data and the directional information D1, a set of values T for parameters of the transfer function, and that set of values T is fed to the transmission unit 20. The transfer function G depends on one or more transfer function parameters. This allows to adjust the transfer function G by assigning different values to at least a part of these transfer function parameters.
In the evaluating unit 50, a link between the audio signals S1 (and, accordingly, the picked-up incoming acoustic sound 6) and the directional information D1 is generated, which is very valuable for assigning such values T to parameters of the transfer function G, which result in an optimized hearing sensation for the user in the current acoustical environment.
The storage unit 60 is optional and may, e.g., be realized in form of some computer memory. The evaluating unit 50 might as well receive the directional information D1 from elsewhere, e.g., from the input unit 10. The directional information D1 is or comprises data related to a directional characteristic, with which the audio signals S1 have been obtained (by means of the input unit 10) from the incoming acoustic sound 6. It may, e.g., comprise data related to a head-related transfer function (HRTF) of the user and/or data related to polar patterns of employed microphones.
In all block-diagrammatical Figures, bold solid arrows depict audio signals, whereas thin solid arrows depict data or control signals.
The input unit 10 comprises at least two input transducers M1,M2 (e.g., microphones), which derive raw audio signals R1 and R2, respectively, from incoming acoustic sound (not depicted in
As symbolized by switch 14, one of the raw audio signals R1,R2 can be selected as audio signal S1 or S2, respectively, and fed to the characterizing unit 40. I.e., the switch 14 symbolizes or indicates a successive (consecutive) obtaining, with different directional characteristics, of audio signals from acoustic sound. The characterization thereof will then usually take place successively.
It is possible to feed said raw audio signals R1,R2 and/or said audio signal S1 or S2, respectively, to the transmission unit 20.
The characterizing unit 40 comprises a feature extractor FE1 and a classifier CLF1. The feature extractor FE1 extracts features f1a,f1b,f1c from the fed-in audio signal S1, and features f2a,f2b,f2c from the fed-in audio signal S2, respectively. These sets of features, which in general may comprise one, two or more (maybe even of the order of ten or 40) features, are fed to classifier CLF1, in which it is classified into one or a number of several possible classes. The classification result is the sound-characterizing data C1 and C2, respectively, or is comprised therein.
For deriving at least a part of the directional information D1, the evaluating unit 50 is operationally connected to the switch 14. Accordingly, the evaluating unit 50 “knows” whether a currently received set of sound-characterizing data is obtained from acoustic sound picked-up with transducer M1 or with transducer M2. Besides the information, with which of the transducers (M1 or M2) acoustic sound has been picked up, the evaluating unit 50 preferably shall also have information about the directional characteristic assigned to the corresponding transducers. Such information (e.g., on HRTFs and polar patterns) may be obtained from the position of switch 14 or from a storage modul in the hearing system (not shown).
The embodiment of
Both raw audio signals R1,R2 will usually be fed also to the transmission unit 20. Additionally or alternatively, said audio signals S1 can be fed to the transmission unit 20, too.
By means of the beam former controller BFC1, the beam former can be adjusted to form a desired directional characteristic, i.e., the directional characteristic is set by means of the beam former. Data related to that desired directional characteristic are at least a part of the directional information D1 and can be transmitted from the beam former controller BFC1 to the evaluation unit 50.
Usually, the beam former will have a preferred direction, i.e., it will be adjusted such that acoustic sound impinging on the transducers M1,M2 from that preferred direction (or angular range) is picked-up with relatively high sensitivity, while acoustic sound from other directions is damped.
It is possible to control the beam former such that only sound from a narrow angular range around the preferred direction is picked up and characterized, and the corresponding sound-characterizing data C1 are then, together with the directional information D1, evaluated, and the transfer function G is thereupon adjusted. Characterization may, e.g., take place by feature extraction and classification.
It is also possible to control the beam former such that first, a first preferred direction (or, more general, a first directional characteristic) is selected, and then a second preferred direction (or, more general, a second directional characteristic) is selected; and optionally after that even more, one after each other. Preferably, a common evaluation of the (at least) two corresponding sets of sound-characterizing data and the corresponding directional information will take place.
In case of two such preferred directions, approximately opposite directions can be chosen. This will usually maximize the information derivable from the common evaluation. For example, the front hemisphere and the back hemisphere can be chosen.
This effect can be considered, and accordingly corrected polar patterns P1,P2 can be obtained by making use of a head-related transfer function (HRTF).
The term head-related transfer function (HRTF) in this application comprises, of course, also approximations of HRTFs, and HRTFs reduced to its relevant parts, e.g., parts considering only the amplitude part of the HRTF and leaving out phase information.
The two microphones M1,M2 (or corresponding microphone arrangements) may be worn on the same side of the user's head or on opposite sides.
It is also possible to control the beam former such that the acoustic environment is investigated in four quadrants, preferably with center directions at approximately 0°, 90°, 180°, 270°. This can be accomplished by simultaneously or successively adjusting the beam former such, that sound originating from a location in 0°, 90°, 180° and 270°, respectively, is amplified stronger or attenuated less than sound originating from other locations. The corresponding four sets of sound-characterizing data can, e.g., be deduced from the four corresponding beam former settings. An evaluation of the corresponding four sets of sound-characterizing data together with their corresponding directional information is preferred.
Another possibility is, to control the beam former such that the acoustic environment is investigated in even more sections.
In
An evaluation of the corresponding (at least) nine audio signals (together with corresponding directional information on each) will give rather deep insight into the location of sources of sound in the surroundings of the user. Accordingly, the transfer function can be adjusted in a way that very well suits the user's needs in that particular situation.
It is possible to realize embodiments as discussed in conjunction with
For optimizing beam former settings, it can be advantageous to introduce a data communication from the evaluating unit 50 to the beam former controller BFC1 (feedback; not shown in
In
Raw audio signals R1,R2,R3 from the input transducers M1,M2,M3, respectively, (or from audio signals derived therefrom) are fed to the localizer L1. Therefrom, the localizer L1 derives that (in this example) three main sources of acoustic sound Q1,Q2,Q3 exist, which are located at polar angles of about 110°, 190° and 330°, respectively.
This information is fed to the evaluation unit 50 as directional informations D1,D2,D3 (or as a part of that), and one beam former each is instructed with information to focus into one of these preferred directions. Accordingly, first, second and third audio signals S1, S2 and S3, respectively, are generated such, that they preferably contain acoustic sound stemming from one of the main sources of acoustic sound Q1, Q2 and Q3, respectively. These audio signals S1, S2 and S3 are separately characterized, in this example by feature extraction and classifying.
In
Each classification result (corresponding to sound-characterizing data) may comprise similarity values indicative of the likeness of the current acoustical scene and an acoustic scene represented by a certain class (“mixed-mode” classification), as shown in
Thus, the link between the knowledge obtained from the localizer, that some sources of acoustic sound are present in the above-mentioned three main directions, and the findings, obtained from the characterizing units (feature extractors and classifiers), about what kind of sound source is apparently located in the respective direction, can be made in the evaluation unit 50. This way, the acoustic environment can be captured rather precisely.
Assuming that, when close to the straight-ahead direction (θ=0°) a speaker (source of a speech signal) exists, the user prefers to understand that speech and wants other signals (like noise and music) to be fully or partially suppressed or muted, a transfer function G (or hearing program) accomplishing this task can be selected. In the current example, the transfer function G may use a beam former, which is adjusted such that acoustic sound impinging on the microphones from θ=110° is suppressed (has low amplification) as far as possible, while acoustic sound from θ=330° is emphasized (has stronger amplification), and acoustic sound from θ=190° is to some extent tolerated.
In this example, the resulting transfer function is possibly not strongly different from what is obtained from a simple classifier-beamformer approach, in which, without the evaluation according to the invention, it would be assumed that in a speech-in-noise situation—if a classification based on not or hardly focussed acoustic signals derives this classification result—the speaker is typically located near θ=0°. In such a simple classifier-beamformer approach, a beam former might be used with a maximum amplification at θ=0°, which probably would let through the speech and suppress the music (190°) well and would provide for some suppression of the noise (110°), too.
By means of the invention, be it using a localizer or using section-wise environment sound investigation or others, it is probably possible to recognize that the three persons A1, A2, A3 exist, and approximately where they are located, and where the noise source N is located, so that the angular range depicted as Δθ (in solid lines) could be selected. Good noise suppression and good intelligibility of the speaker will be achieved.
And, as has already been described above, it is also possible to have, for determining the set of values T for transfer function parameters, only one beam former unit and one characterizing unit, which process audio signals obtained from acoustic sound, one after the other, with different directional characteristics.
The output unit 80 may have one or two output transducers (e.g., loudspeakers or implanted electrical-to-electrical or electrical-to-mechanical converters). If two output transducers are present, these will typically be fed with two different (partial) output audio signals 7.
From signals S1 and S2, respectively, which are obtained from the input transducers M1 and M2, respectively, sets of features are extracted and classified. In
Preferably, a “mixed-mode” classification (described above) is used. From the so-obtained similarity vectors (embodying sound-characterizing data C1,C2), in conjunction with directional information D1,D2, information about the location (direction) of the speech source and of the noise source may be derived. The directional information D1,D2 may comprise HRTF-information and/or information on the directional characteristics of the microphones M1,M2, preferably both (which would approximately correspond to experimentally determined directional characteristics when the hearing system is worn, at the user or at a dummy).
The evaluation may take place in one of the two hearing devices, in which case at least one of the sets C1,C2 of sound-characterizing data has to be transmitted from one hearing device to the other. Or the evaluation may take place in both hearing devices, in which case the sets C1,C2 of sound-characterizing data have to be interchanged between the two hearing devices. It would also be possible to do the feature extraction and classification in only one of the hearing devices, in which case the audio signals S1 or S2 have to be transmitted to from one hearing device to the other.
The transmission unit 20 and transfer function G may be realized in one or in both hearing devices, and it may process audio data for one or in both hearing devices. For example, the hearing system might be a cross-link hearing system, which picks-up acoustic sound on both sides of the head, but outputs sound only on one side.
In a binaural system, it can be decided, whether the sound characterization and/or the evaluation and/or the transfer function processing shall take place in one or both of the hearing devices. Therefrom results the necessity to transmit input audio signals, sound-characterizing data, sets of values for transfer function parameters of (partial) transfer functions and/or (partial) output audio signals from one of the two hearing devices to the other.
In general, it has to be noted that throughout the text above, details of the transfer functions and their parameters have only been roughly discussed, because a major aspect of the invention is related to ways for obtaining values for transfer function paramters. Often, it will be advantageous to provide for a beam forming function within the transfer function. Such a beam former may use the same settings as a beam former, which is possibly used for deriving audio signals, which are to be characerized in order to derive sound-characterizing data for the evaluation unit. But different settings may be used as well. The same physical beam former may be used for both tasks, or different ones, and beam formers may be realized in form of software, so that various beam former software modules may run in parallel or successively for finding values for transfer function parameters and for the transfer function itself, i.e., for signal processing in the transmission unit.
In embodiments described above, at least one pair of data comprising
Number | Date | Country | |
---|---|---|---|
60747345 | May 2006 | US |