The present disclosure deals with a body worn hearing system, e.g. a hearing aid system. The hearing system comprises a hearing device, or a pair of hearing devices (e.g. hearing aids), and a separate microphone unit.
The present disclosure relates in particular to a hearing system configured to be used by a hearing impaired person (‘the user’) and comprising a separate microphone unit, e.g. in the form or a wireless, e.g. clip-on-, microphone unit, which may be used to transmit a user's own voice to a communication device, e.g. a telephone (such as a cellular telephone). Such a microphone unit may comprise an array of M microphones (i.e. M≧2), which by use of (e.g. adaptive) beamforming may enhance the voice of the person talking. Our co-pending European patent application no. EP16154471.3, filed with the EPO on 5 Feb. 2016, and published as EP3057337A1 deals with the same topic. In EP3057337A1, it is proposed to build a dedicated adaptive beamformer and single-channel noise reduction (SC-NR) algorithm into the separate microphone unit, which in a specific communication (e.g. telephone reception) situation is able to retrieve a voice signal of the user wearing the microphone unit from the noisy microphone signals received by the microphone unit, and to reject/suppress other sound sources.
A Hearing System:
The present disclosure proposes a number of features that can be used to improve a body worn hearing system in a communication mode, where a wearer's own voice is picked up (by a separate microphone unit) and transmitted to another device (a communication device).
As the person talking (e.g. the mouth of the person wearing the microphone unit) is close to the microphone unit, the sound of interest is in the acoustic near-field. When the sound of interest is in the near field, the sound pressure level at the (e.g. two) microphones may differ because one microphone is further away from the mouth compared to the other(s). The difference in sound pressure level will depend on the distance between the mouth and the microphone unit. If the microphone unit is relatively close to the mouth, the sound pressure level difference will be higher compared to the sound pressure level difference if the microphone unit is relatively further away from the mouth. When the sound is in the far-field, the relative distance between the microphones compared to the distance between the microphones and the sound source, becomes small, and the difference in sound pressure level between the microphones becomes insignificant. For near-field applications, in order to achieve an optimal directional response, we need to take into account that the transfer function (or impulse response) between the microphones not only depends on the direction to the sound source but also on the distance to the sound source, cf.
It is an object of the present disclosure to provide an alternative directional microphone system for picking up a user's voice in a body-worn hearing system. It is an object of an embodiment of a hearing system according to the present disclosure to provide a scheme for estimating a distance, or a propagation time delay, or (relative) transfer functions between the mouth of a user and microphones of a microphone unit and/or an angle between a reference direction of the microphone unit and a direction to the mouth of the user (when mounted on the body of the user). It is a further object to provide that parameters related to (e.g. dependent on) a current geometric configuration of the microphone unit relative to the mouth of the user (e.g. relative transfer functions RTFs, distances D, time delays Δt, or tilt angle θ) are updated at appropriate points in time (acoustic conditions) and used to control a noise reduction system (e.g. a beamformer filtering unit), at least in a specific communication mode of operation of the hearing system.
In an aspect of the present application, a body worn hearing system comprising a hearing device, e.g. a hearing aid, adapted for being located at or in an ear of a user, or adapted for being fully or partially implanted in the head of the user, and a separate microphone unit adapted for being located at said user and picking up a sound, e.g. a voice of the user, from the user's mouth, is provided.
The hearing device comprises
The microphone unit comprises
The hearing system comprises a control unit configured to estimate
The hearing system is configured to control the multi-input noise reduction system in dependence of said current distance or said current time delay or said relative transfer functions.
Thereby an improved hearing system may be provided.
By estimating
a set of beamformer weights of the beamformer filtering unit can be adaptively updated, e.g. by selecting an appropriate set of beamformer weights from a number of sets of beamformer weights (w(D (or Δt, or RTF), θ, k), k being a frequency index, k=1, . . . , K, where K is the number of frequency sub-bands). The data constitute a dictionary of beamformer weights corresponding to specific different values of distance D, or propagation time delay Δt, or relative transfer functions RTF (and possibly angle θ), are e.g. stored in a memory of the hearing system (or accessible to the hearing system).
In an embodiment, the dictionary comprises corresponding values of:
where k=1, . . . , K.
In an embodiment, the dictionary comprises values of relative transfer functions RTFp(D, θ, k) instead of, or in addition to, the beamformer weights w(D, θ, k).
In an embodiment, the dictionary comprises corresponding (e.g. predetermined) values of distance (or time delay), or relative transfer functions, and beamformer filtering weights for a number of different locations of the target sound source relative to the microphone unit, e.g. including the user's mouth, and one or more of a table and another person. In an embodiment, current estimates of the distance (or time delay) or relative transfer functions are used to determine where the microphone is located.
When tilt angle is included, a set of frequency dependent beamformer weights w(k) for each distance D (or time delay Δt, or relative transfer functions RTF) and each tilt angle θ is available in the dictionary (or database), i.e. in total ND times Nθ sets of beamformer weights w(k). In an embodiment, such sets of beamformer weights are determined in advance of operation of the hearing system and stored on a medium accessible to the hearing system, e.g. in a memory of the microphone unit.
The distance D and propagation time delay Δt is tied together by the velocity of sound. For propagation in air, D(MO-Mi)=cair·Δt(MO-Mi), where the ‘variable’ MO-Mi represent a specific configuration of audio source (mouth, MO) and microphone (Mi, i=1, . . . , M).
The spatially filtered signal from the beamformer filtering unit may be equal to the estimate of the target signal s comprising the user's voice. In an embodiment the spatially filtered signal is further processed (e.g. in a single channel noise reduction unit or other post-processing unit) to provide the estimate Ŝ of the target signal s (cf. e.g.
In an embodiment, the control unit is configured to estimate a current distance or a current time delay (and/or relative transfer functions) from the user's mouth to the at least one, such as a majority or all, of the multitude M of microphones of the microphone unit. In an embodiment, the geometrical configuration of the multitude M of microphones Mi, i=1, 2, . . . , M, is known (e.g. fixed within the microphone unit). In an embodiment, (at least some of, such as all of) the mutual distances Lij between the microphones are known (i=1, 2, . . . . , M, j=1, 2, . . . , M, while i≠j), and e.g. stored in a memory of the hearing system (or accessible to the hearing system). In an embodiment, the microphones are located on one straight line. In an embodiment, Lij=L for all j=i+1, i=1, 2, . . . , M−1. In an embodiment, M=2. In an embodiment, M=3. In an embodiment, M=4.
The term ‘a tilt angle θ of the microphone unit’ is in the presence context taken to mean an angle θ defined by the microphone unit (e.g. its housing, or a feature of the housing, e.g. an imprint or a mechanical protrusion or indentation, or any other characteristic feature of the microphone unit defining an axis) and a reference direction (e.g. a direction of the acceleration of gravity).
In an embodiment, the microphone unit comprises a housing wherein or whereon the multitude M of microphones are located, the housing defining a microphone unit reference direction MDREF. In an embodiment, the microphone unit reference direction MDREF is defined by or related to an edge or surface of the housing of the microphone unit. In an embodiment, the microphone unit reference direction MDREF is defined by or related to a geometrical configuration of the multitude M of microphones. In an embodiment the microphone unit reference direction MDREF is defined by or related to a microphone direction defined by two of the microphones of the multitude M of microphones (e.g. by a straight line through the two microphones). In an embodiment, the orientation of the microphone unit relative to a direction from the microphone unit to the user's mouth is defined by an angle between the microphone unit reference direction MDREF and the direction MO-MD from the microphone unit to the user's mouth.
In an embodiment, the antenna and transceiver unit of the hearing device comprises separate first and second antenna and transceiver units, wherein
In an embodiment, the first antenna and transceiver unit of the hearing device is configured to establish the communication link to the communication device and to additionally transmit information to the communication device, at least in a specific communication mode of operation of the hearing system.
In an embodiment, the antenna and transceiver unit of the microphone unit comprises separate first and second antenna and transceiver units, wherein
In an embodiment, the control unit is configured to estimate a current orientation of the microphone unit relative to a direction from the microphone unit to the user's mouth, and wherein the hearing system is configured to control the multi-input noise reduction system in dependence of the orientation of the microphone unit relative to a direction from the microphone unit to the user's mouth. If the microphone unit is tilted (so that a reference direction MDREF of the microphone unit (e.g. an axis between two microphones) is not pointing in the direction of the mouth of the user), see e.g.
The look vector d (RTF) is in the present context taken to be a representation of a normalized acoustic transfer function from a target sound source (at a given location, here the user's own voice, i.e. from the mouth of the user) to each microphone Mi, i=1, . . . , M, of the microphone unit, i.e. d is an M dimensional vector. In an embodiment, d=d′/SQRT(|d′|2), where d′ is the un-normalized look vector.
In an embodiment, the input unit is configured to provide said time varying electric inputs signals x′i(n) as electric input signals Xi(k,m) in a time-frequency representation comprising time varying signals in a number of frequency sub-bands, k being a frequency band index, m being a time index. In an embodiment, m is a time-frame index. In an embodiment, the multi-input noise reduction system is configured to determine filter weights w(k,m) for providing the spatially filtered (‘beamformed’) signal, wherein signal components from other directions than a direction of a target signal source are attenuated, whereas signal components from the direction of the target signal source are left un-attenuated or are attenuated less relative to signal components from said other directions. In an embodiment, the current distance (or delay or relative transfer functions) (at time m′) is used to select appropriate beamformer filter weights w(k,m′).
In an embodiment, the multi-input beamformer filtering unit is configured to be adaptive.
In an embodiment, a transfer function (and/or relative transfer function) from the target sound source (the user's mouth) to a microphone of the microphone unit is determined while the user is talking The transfer function may e.g. be determined when the hearing system is in a communication mode, e.g. during a telephone conversation, where a two-way (bi-directional) link to a ‘far-end person’ is established via a telephone and a telephone network (e.g. the Internet and/or via a public switched telephone network (PSTN)). In such situation, the user is likely to talk, if the far-away line from the ‘far-end person’ is quiet. In an embodiment, at least one of the left and right hearing devices HDL and HDR, are configured to receive a direct electric audio signal from a telephone (representing the voice of the far-end communication partner). In an embodiment, at least one of the left and right hearing devices comprises a voice or speech activity detector for determining whether (or with which probability) a voice is present in the received direct electric audio signal (the telephone signal). In an embodiment, the microphone unit, and/or the hearing device comprises an own voice detector for estimating whether (or with which probability) a user's own voice is present in the microphone signals picked up by the microphone unit and/or the hearing device. In an embodiment, the transfer functions (and/or relative transfer functions, or delay or distance) are estimated on initiation of a user (or as a standard procedure during power-on of the hearing system), e.g. via a user interface, e.g. under the condition that a detected environment sound level is below a threshold level (whereby a high SNR during estimation can be obtained). In an embodiment, an activation of the estimation of the transfer functions, etc., is indicated to the user via a loudspeaker of the hearing device(s) as an acoustic invitation to the user to speak, e.g. a predefined word or words or sentence(s), see e.g.
In an embodiment, the hearing device comprises a voice activity or speech detector configured to determining whether, or with which probability, a voice is present in the direct electric audio signal received from the communication device.
In an embodiment, the microphone unit comprises a voice or speech activity detector configured to determining whether, or with which probability, a voice or speech (in particular a user's own voice or speech) is present in the spatially filtered signal or in one or more of the electric input signals representative of sound from the environment of the microphone unit.
Voice activity information can e.g. be used to provide an adaptive noise reduction system (NRS) to control the timing of the update of the noise reduction system (e.g. update a look vector d when the user speaks (target sound source (=user's voice) is present), and update a noise covariance matrix Rvv when the user does NOT speak. In an embodiment, this is indicated by a detection of no voice activity (update of d, assuming that the hearing system user speaks) and of voice/speech activity (update of Rvv, assuming that the hearing system user does not speak) in the wirelessly received signal (i.e. from a speaker at ‘the other end’ of the telephone line).
In an embodiment, the hearing system comprises a single hearing device (only one). In an embodiment, the input unit of the hearing device comprises at least one microphone for converting a sound from the environment to an electric input signal. In an embodiment, the hearing system comprises left and right hearing devices, e.g. hearing aids, adapted for being located at or in respective left and right ear of a user, or adapted for being fully or partially implanted in the head at respective left and right ear of a user. In normal use, the distances DL and/or DR between the mouth and the left and right hearing devices HDL and HDR, respectively, do typically not vary much from day to day (where the hearing instruments have been dismounted and mounted again). But the distance between the body worn microphone unit and the mouth may be different every time the body worn microphone unit is mounted (the body worn microphone unit being a separate unit, that is mounted on the user's body (e.g. attached to the body, e.g. to clothing) independently of the hearing device(s)). In an embodiment, the hearing system comprises a detection unit configured to detect a difference in acoustic propagation time between sound from the user's mouth to the hearing device and to the microphone unit, respectively. In an embodiment, the detection unit is configured to detect a difference in acoustic propagation time between sound from the user's mouth to the at least one microphone of the hearing device and to one of the multitude M of microphones of the microphone unit, respectively. In an embodiment, the detection unit is configured to determine a similarity, e.g. a correlation, such as a cross correlation, between sound from the user's mouth received at the hearing device and at the microphone unit.
In an embodiment, the detection unit is configured to determine a cross correlation between sound from the user's mouth received at a microphone of the hearing device and sound received at one of the multitude M of microphones of the microphone unit. The cross-correlation is used to determine a difference in time of arrival (ta) of acoustic signals from the user's mouth to the respective microphones (thereby identifying the time difference Δt(HD-MICU)=ta(HD)−ta(MICU) that provides an optimal value of the cross-correlation. Knowing the distance DR (and/or DL) between the user's mouth and the hearing device (HDR, HDL) see
In an embodiment, the microphone unit and/or the hearing device is/are configured to align the electric signals from the microphones in time, so that a given acoustic event (e.g. speech) is provided in the (aligned) signal streams at the same time. Thereby any given microphone signal may be selected for processing and/or presentation to the user (or a communication partner) at a given time in dependence of a current application or acoustic situation (e.g. own voice, acoustic feedback, reverberation, etc.).
In an embodiment, the hearing device(s) and the microphone unit are adapted to establish a communication link between them allowing information signals to be exchanged between them. In an embodiment, the information signals include audio signals (or parts of audio signals, e.g. selected frequency bands), and/or one or more parameters related to the current distance and/or direction from the user's mouth to the microphone unit, and/or current relative transfer functions from the user's mouth to the individual microphones of the microphone unit.
In an embodiment, the antenna and transceiver units of the hearing device and the microphone unit each comprises respective antenna coils configured to have an inductive coupling to each other that allow an inductive communication link to be established between the hearing device and the microphone unit when the hearing device and the microphone unit are mounted on the user's body, and wherein at least one of the hearing device and the microphone unit comprises at least two mutually angled antenna coils. In an embodiment, each antenna coil exhibits a coil axis defined by a center axis of a (virtual or physical) carrier around which the winding of the coil extends (around which the turns are wound). In an embodiment, the antenna and transceiver unit of the microphone unit comprises two or three mutually angled (e.g. orthogonal) antenna coils (in other words, the axes of the antenna coils are angled). In an embodiment, the antenna and transceiver unit of the hearing device comprises a single antenna coil. Assuming that the orientation of the coil-axis (or axes) of the antenna coil(s) of the hearing device when mounted on the user U is known and assuming that the orientation of the antenna coils of the microphone unit relative to the reference axis of the microphone unit is known, an orientation of the microphone unit relative to a global reference direction GDREF (e.g. the direction of the force of gravity) can be determined based on the relative signal strengths of the electromagnetic signals received at the respective antenna coils of the microphone unit from the antenna coil of the hearing device.
In an embodiment, the hearing system is configured to be able to access a dictionary of beamformer weights (and/or relative transfer functions) and corresponding mouth to microphone unit distances, or time delays and optionally tilt angles. In an embodiment, the hearing system comprises a memory wherein a dictionary of beamformer weights and corresponding mouth to microphone distances, or time delays, or relative transfer functions, and optionally tilt angles is stored. In an embodiment, the microphone unit comprises said memory. In an embodiment, a number of different sets of beamformer weights corresponding to different distances Dp, p=1, . . . , ND, between the mouth of the user and the microphone unit are stored in the memory. In this way, an appropriate set of beamformer weights can be chosen and applied, when a current distance (and/or direction (e.g. angle θq, q=1, . . . , Nθ)) has been determined. In an embodiment, a number of different sets of beamformer weights corresponding to different distances Dp, p=1, . . . , ND, (or corresponding propagation time delays Δtp) between the mouth of the user and the microphone unit, or relative transfer functions (RTFp), are stored in the memory together with a number of tilt angles θq, q=1, . . . , Nθ of the microphone unit for each distance Dp (or Δtp, or RTFp). In an embodiment, a current distance D′, or a current propagation time delay Δt′, or current relative transfer functions RTF′ is/are determined by the hearing system. In an embodiment, the hearing system is configured to select an appropriate set of beamformer weights (and or relative transfer functions)by testing each of the stored sets of beamformer weights w(Dp (or Δtp), k) and/or relative transfer functions RTFp(D, θ, k) corresponding to different tilt angles θq, q=1, . . . , Nθ, k=1, . . . , K, and to choose the set that optimizes the user's voice (e.g. using a maximum likelihood framework, where a likelihood function is determined and an estimated distance or tilt angle or beamformer weights is selected as the one corresponding to a maximum value of the likelihood function, cf. e.g. [Farmani et al.; 2017]).
In an embodiment, the (body worn) hearing system comprises a pair of hearing devices (e.g. denoted first and second hearing devices), e.g. adapted for being located at (or fully or partially implanted in) left and right ears, respectively, of a user. In an embodiment, the pair of hearing devices form part of a binaural hearing system, e.g. a binaural hearing aid system. In an embodiment, the left and right hearing devices each comprises antenna and transceiver circuitry allowing the exchange of information between them. In an embodiment, such information may comprise audio data and/or control signals and/or status signals.
In an embodiment, the (or each) hearing device comprises a hearing aid.
In an embodiment, the hearing system comprises a user interface allowing a user to influence functionality of the system. The user interface may be implemented fully or partially in the hearing device or in the microphone unit, or in an auxiliary device. In an embodiment, the hearing system is configured to allow an initiation of a procedure for updating current values of a look vector (RTFs, or distance D, or delay Δt).
In an embodiment, the hearing system comprises an auxiliary device implementing a user interface for the hearing system. In an embodiment, the auxiliary device (and the user interface) is configured to allow the exchange of information with the hearing system (e.g. the hearing device, and/or the microphone unit) via appropriate communication links. The user interface is preferably configured to allow a user to influence functionality of the hearing system, e.g. to enter or leave a specific communication mode according to the present disclosure. In an embodiment, the user interface is configured to allow a user to initiate an update of parameters (e.g. RTFs, D, Δt) related to (e.g. dependent on) a current geometric configuration of the microphone unit relative to the mouth of the user. The user interface may further be configured to allow presentation of information to a status of the hearing system. In an embodiment, the auxiliary device comprises a remote control device, e.g. a smartphone. In an embodiment, the user interface is implemented as an APP of a smartphone.
In an embodiment, the microphone unit is implemented in the auxiliary device together with a user interface, e.g. in smartphone.
In an embodiment, the hearing system is configured to initiate an update of parameters related to (e.g. dependent on) a current geometric configuration of the microphone unit relative to the mouth of the user (e.g. RTFs, D, Δt) during start-up (e.g. power on) of the system, during use (e.g. when specific criteria or acoustic conditions are fulfilled), or continuously, or allowing such initiation to be performed via a user interface (via a ‘user speech test’). The user interface may e.g. be implemented as a button on the hearing device, or as a remote control device (e.g. comprising an interactive display), or form part of an APP running on a cellular phone, e.g. a smartphone, a smartwatch, or similar portable or wearable device.
A Hearing Device:
In an aspect, a hearing device, e.g. a hearing aid, adapted for being located at or in an ear of a user, or adapted for being fully or partially implanted in the head of the user, is further provided by the present disclosure. The hearing aid is configured to form part of a hearing system as described above, in the detailed description of embodiments, and in the claims.
In an embodiment, the hearing device is adapted to provide a frequency dependent gain and/or a level dependent compression and/or a transposition (with or without frequency compression) of one or frequency ranges to one or more other frequency ranges, e.g. to compensate for a hearing impairment of a user. In an embodiment, the hearing device comprises a signal processing unit for enhancing the input signals and providing a processed output signal.
In an embodiment, the hearing device comprises an output unit for providing a stimulus perceived by the user as an acoustic signal based on a processed electric signal. In an embodiment, the output unit comprises a number of electrodes of a cochlear implant or a vibrator of a bone conducting hearing device. In an embodiment, the output unit comprises an output transducer. In an embodiment, the output transducer comprises a receiver (loudspeaker) for providing the stimulus as an acoustic signal to the user. In an embodiment, the output transducer comprises a vibrator for providing the stimulus as mechanical vibration of a skull bone to the user (e.g. in a bone-attached or bone-anchored hearing device).
In an embodiment, the hearing device comprises an input unit for providing an electric input signal representing sound. In an embodiment, the input unit comprises an input transducer, e.g. a microphone, for converting an input sound to an electric input signal. In an embodiment, the input unit comprises a wireless receiver for receiving a wireless signal comprising sound and for providing an electric input signal representing said sound. In an embodiment, the hearing device comprises a directional microphone system adapted to spatially filter sounds from the environment, and thereby enhance a target acoustic source among a multitude of acoustic sources in the local environment of the user wearing the hearing device. In an embodiment, the directional system is adapted to detect (such as adaptively detect) from which direction a particular part of the microphone signal originates. This can be achieved in various different ways as e.g. described in the prior art.
In an embodiment, the hearing device comprises an antenna and transceiver circuitry for establishing a wireless link to (e.g. wirelessly receiving a direct electric input signal from) another device, e.g. a communication device or another hearing device. In an embodiment, the communication between the hearing device and the other device is based on some sort of modulation at frequencies above 100 kHz. Preferably, frequencies used to establish a communication link between the hearing device and the other device is below 70 GHz, e.g. located in a range from 100 kHz to 50 MHz, or in a range from 50 MHz to 70 GHz, e.g. above 300 MHz, e.g. in an ISM range above 300 MHz, e.g. in the 900 MHz range or in the 2.4 GHz range or in the 5.8 GHz range or in the 60 GHz range (ISM=Industrial, Scientific and Medical, such standardized ranges being e.g. defined by the International Telecommunication Union, ITU). In an embodiment, the wireless link is based on a standardized or proprietary technology. In an embodiment, the wireless link is based on Bluetooth technology (e.g. Bluetooth Low-Energy technology). In an embodiment, the wireless link is based on near-field communication, e.g. inductive communication.
In an embodiment, the hearing device is a portable device, e.g. a device comprising a local energy source, e.g. a battery, e.g. a rechargeable battery.
In an embodiment, the hearing device comprises a forward or signal path between an input transducer (microphone system and/or direct electric input (e.g. a wireless receiver)) and an output transducer. In an embodiment, the signal processing unit is located in the forward path. In an embodiment, the signal processing unit is adapted to provide a frequency dependent gain according to a user's particular needs. In an embodiment, the hearing device comprises an analysis path comprising functional components for analyzing the input signal (e.g. determining a level, a modulation, a type of signal, an acoustic feedback estimate, etc.). In an embodiment, some or all signal processing of the analysis path and/or the signal path is conducted in the frequency domain. In an embodiment, some or all signal processing of the analysis path and/or the signal path is conducted in the time domain.
In an embodiment, an analogue electric signal representing an acoustic signal is converted to a digital audio signal in an analogue-to-digital (AD) conversion process, where the analogue signal is sampled with a predefined sampling frequency or rate fs, fs being e.g. in the range from 8 kHz to 48 kHz (adapted to the particular needs of the application) to provide digital samples xn (or x[n]) at discrete points in time tn (or n), each audio sample representing the value of the acoustic signal at tn by a predefined number Ns of bits, Ns being e.g. in the range from 1 to 48 bits, e.g. 24 bits. A digital sample x has a length in time of 1/fs, e.g. 50 μs, for fs=20 kHz. In an embodiment, a number of audio samples are arranged in a time frame. In an embodiment, a time frame comprises 64 or 128 audio data samples. Other frame lengths may be used depending on the practical application.
In an embodiment, the hearing devices comprise an analogue-to-digital (AD) converter to digitize an analogue input with a predefined sampling rate, e.g. 20 or 24 or 32 or 48 kHz. In an embodiment, the hearing devices comprise a digital-to-analogue (DA) converter to convert a digital signal to an analogue output signal, e.g. for being presented to a user via an output transducer.
In an embodiment, the hearing device, e.g. the microphone unit, and or the transceiver unit comprise(s) a TF-conversion unit for providing a time-frequency representation of an input signal. In an embodiment, the time-frequency representation comprises an array or map of corresponding complex or real values of the signal in question in a particular time and frequency range. In an embodiment, the TF conversion unit comprises a filter bank for filtering a (time varying) input signal and providing a number of (time varying) output signals each comprising a distinct frequency range of the input signal. In an embodiment, the TF conversion unit comprises a Fourier transformation unit for converting a time variant input signal to a (time variant) signal in the frequency domain. In an embodiment, the frequency range considered by the hearing device from a minimum frequency fmin to a maximum frequency fmax comprises a part of the typical human audible frequency range from 20 Hz to 20 kHz, e.g. a part of the range from 20 Hz to 12 kHz. In an embodiment, a signal of the forward and/or analysis path of the hearing device is split into a number NI of frequency bands, where NI is e.g. larger than 5, such as larger than 10, such as larger than 50, such as larger than 100, such as larger than 500, at least some of which are processed individually. In an embodiment, the hearing device is/are adapted to process a signal of the forward and/or analysis path in a number NP of different frequency channels (NP≦NI). The frequency channels may be uniform or non-uniform in width (e.g. increasing in width with frequency), overlapping or non-overlapping.
In an embodiment, the hearing device (and/or the microphone unit) comprises a number of detectors configured to provide status signals relating to a current physical environment of the hearing device (e.g. the current acoustic environment), and/or to a current state of the user wearing the hearing device, and/or to a current state or mode of operation of the hearing device. Alternatively or additionally, one or more detectors may form part of an external device in communication (e.g. wirelessly) with the hearing device. An external device may e.g. comprise another hearing assistance device, a remote control, the microphone unit, an audio delivery device, a telephone (e.g. a Smartphone), an external sensor, etc.
In an embodiment, one or more of the number of detectors operate(s) on the full band signal (time domain). In an embodiment, one or more of the number of detectors operate(s) on band split signals ((time-) frequency domain).
In an embodiment, the number of detectors comprises a level detector for estimating a current level of a signal of the forward path. In an embodiment, the predefined criterion comprises whether the current level of a signal of the forward path is above or below a given (L-)threshold value. In an embodiment, the level detector or a control unit connected to the level detector is configured to estimate whether a current sound level is in a normal range for own voice levels (ΔLov). In an embodiment, the level detector or a control unit connected to the level detector is configured to estimate whether a current sound level is below a predefined (background) threshold level (Lbg), where it can be assumed that the user's own voice is NOT present. The normal range for own voice levels (ΔLov) and the predefined (background) threshold level (Lbg) may e.g. be predefined, e.g. measured or estimated in advance of (normal) use of the hearing system, E.g. stored in a memory of the hearing system (or accessible to the hearing system).
In an embodiment, the hearing system is configured to (automatically) estimate the parameters related to a current geometric configuration of the microphone unit relative to the mouth of the user (e.g. RTFs, D, Δt) when the current sound level is in a normal range for own voice levels (ΔLov). In an embodiment, the hearing system is configured to initiate a user speech test (cf. e.g.
In a particular embodiment, the hearing device comprises a voice detector (VD) for determining whether or not an input signal comprises a voice signal (at a given point in time). A voice signal is in the present context taken to include a speech signal from a human being. It may also include other forms of utterances generated by the human speech system (e.g. singing). In an embodiment, the voice detector unit is adapted to classify a current acoustic environment of the user as a VOICE or NO-VOICE environment. This has the advantage that time segments of the electric microphone signal comprising human utterances (e.g. speech) in the user's environment can be identified, and thus separated from time segments only comprising other sound sources (e.g. artificially generated noise).
In an embodiment, the voice detector is adapted to detect as a VOICE also the user's own voice. Alternatively, the voice detector is adapted to exclude a user's own voice from the detection of a VOICE.
In an embodiment, the hearing device comprises an own voice detector for detecting whether a given input sound (e.g. a voice) originates from the voice of the user of the system. In an embodiment, the microphone system of the hearing device is adapted to be able to differentiate between a user's own voice and another person's voice and possibly from NON-voice sounds.
In an embodiment, the hearing assistance device comprises a classification unit configured to classify the current situation based on input signals from (at least some of) the detectors, and possibly other inputs as well. In the present context ‘a current situation’ is taken to be defined by one or more of
a) the physical environment (e.g. including the current electromagnetic environment, e.g. the occurrence of electromagnetic signals (e.g. comprising audio and/or control signals) intended or not intended for reception by the hearing device, or other properties of the current environment than acoustic;
b) the current acoustic situation (input level, feedback, etc.), and
c) the current mode or state of the user (movement, temperature, etc.);
d) the current mode or state of the hearing assistance device (program selected, time elapsed since last user interaction, etc.) and/or of another device in communication with the hearing device (e.g. the microphone unit, and/or an auxiliary device).
In an embodiment, the hearing device further comprises other relevant functionality for the application in question, e.g. compression, feedback suppression, active noise cancellation, etc.
In an embodiment, the hearing device comprises a hearing aid, e.g. a hearing instrument, e.g. a hearing instrument adapted for being located at the ear or fully or partially in the ear canal of a user. In an embodiment, the hearing device comprises a hearing aid, a headset, an earphone, an ear protection device or a combination thereof.
Use:
In an aspect, use of a hearing system as described above, in the ‘detailed description of embodiments’ and in the claims, is moreover provided. In an embodiment, use is provided in a system comprising audio distribution. In an embodiment, use is provided in a system comprising one or more hearing aids (e.g. hearing instruments, headsets, ear phones, active ear protection systems, etc.), e.g. in handsfree telephone systems, teleconferencing systems, public address systems, karaoke systems, classroom amplification systems, etc.
An APP:
In a further aspect, a non-transitory application, termed an APP, is furthermore provided by the present disclosure. The APP comprises executable instructions configured to be executed on an auxiliary device to implement a user interface for a hearing device or a hearing system described above in the ‘detailed description of embodiments’, and in the claims. In an embodiment, the APP is configured to run on a cellular phone, e.g. a smartphone, or on another portable device allowing communication with said hearing device or said hearing system.
Definitions:
The ‘near-field’ of an acoustic source is a region close to the source where the sound pressure and acoustic particle velocity are not in phase (wave fronts are not parallel). In the near-field, acoustic intensity can vary greatly with distance (compared to the far-field). The near-field is generally taken to be limited to a distance from the source equal to about a wavelength of sound. The wavelength of sound is given by λ=c/f, where c is the speed of sound in air (cair=343 m/s, @ 20° C.) and f is frequency. At f=1 kHz, e.g., the wavelength of sound is 0.343 m (i.e. 34 cm). In the acoustic ‘far-field’, on the other hand, wave fronts are parallel and the sound field intensity decreases by 6 dB each time the distance from the source is doubled (inverse square law).
In the present context, a ‘hearing device’ refers to a device, such as e.g. a hearing aid, or a hearing instrument or an active ear-protection device or other audio processing device, which is adapted to improve, augment and/or protect the hearing capability of a user by receiving acoustic signals from the user's surroundings, generating corresponding audio signals, possibly modifying the audio signals and providing the possibly modified audio signals as audible signals to at least one of the user's ears. A ‘hearing device’ further refers to a device such as an earphone or a headset adapted to receive audio signals electronically, possibly modifying the audio signals and providing the possibly modified audio signals as audible signals to at least one of the user's ears. Such audible signals may e.g. be provided in the form of acoustic signals radiated into the user's outer ears, acoustic signals transferred as mechanical vibrations to the user's inner ears through the bone structure of the user's head and/or through parts of the middle ear as well as electric signals transferred directly or indirectly to the cochlear nerve of the user.
The hearing device may be configured to be worn in any known way, e.g. as a unit arranged behind the ear with a tube leading radiated acoustic signals into the ear canal or with a loudspeaker arranged close to or in the ear canal, as a unit entirely or partly arranged in the pinna and/or in the ear canal, as a unit attached to a fixture implanted into the skull bone, as an entirely or partly implanted unit, etc. The hearing device may comprise a single unit or several units communicating electronically with each other.
More generally, a hearing device comprises an input transducer for receiving an acoustic signal from a user's surroundings and providing a corresponding input audio signal and/or a receiver for electronically (i.e. wired or wirelessly) receiving an input audio signal, a (typically configurable) signal processing circuit for processing the input audio signal and an output means for providing an audible signal to the user in dependence on the processed audio signal. In some hearing devices, an amplifier may constitute the signal processing circuit. The signal processing circuit typically comprises one or more (integrated or separate) memory elements for executing programs and/or for storing parameters used (or potentially used) in the processing and/or for storing information relevant for the function of the hearing device and/or for storing or logging information (e.g. processed information, e.g. provided by the signal processing circuit), e.g. for use in connection with an interface to a user and/or an interface to a programming device. In some hearing devices, the output means may comprise an output transducer, such as e.g. a loudspeaker for providing an air-borne acoustic signal or a vibrator for providing a structure-borne or liquid-borne acoustic signal. In some hearing devices, the output means may comprise one or more output electrodes for providing electric signals.
In some hearing devices, the vibrator may be adapted to provide a structure-borne acoustic signal transcutaneously or percutaneously to the skull bone. In some hearing devices, the vibrator may be implanted in the middle ear and/or in the inner ear. In some hearing devices, the vibrator may be adapted to provide a structure-borne acoustic signal to a middle-ear bone and/or to the cochlea. In some hearing devices, the vibrator may be adapted to provide a liquid-borne acoustic signal to the cochlear liquid, e.g. through the oval window. In some hearing devices, the output electrodes may be implanted in the cochlea or on the inside of the skull bone and may be adapted to provide the electric signals to the hair cells of the cochlea, to one or more hearing nerves, to the auditory brainstem, to the auditory midbrain, to the auditory cortex and/or to other parts of the cerebral cortex.
A ‘hearing system’ refers to a system comprising one or two hearing devices, and a ‘binaural hearing system’ refers to a system comprising two hearing devices and being adapted to cooperatively provide audible signals to both of the user's ears. Hearing systems or binaural hearing systems may further comprise one or more ‘auxiliary devices’, which communicate with the hearing device(s) and affect and/or benefit from the function of the hearing device(s). Auxiliary devices may be e.g. remote controls, audio gateway devices, mobile phones (e.g. SmartPhones), public-address systems, car audio systems or music players. Hearing devices, hearing systems or binaural hearing systems may e.g. be used for compensating for a hearing-impaired person's loss of hearing capability, augmenting or protecting a normal-hearing person's hearing capability and/or conveying electronic audio signals to a person.
Embodiments of the disclosure may e.g. be useful in applications such as hearing aids, headsets, active ear protection systems, or combinations thereof. The disclosure may further be useful in applications combining hearing aids with communication devices, such as headsets, handsfree telephone systems, mobile telephones, teleconferencing systems, public address systems, karaoke systems, classroom amplification systems, etc.
The aspects of the disclosure may be best understood from the following detailed description taken in conjunction with the accompanying figures. The figures are schematic and simplified for clarity, and they just show details to improve the understanding of the claims, while other details are left out. Throughout, the same reference numerals are used for identical or corresponding parts. The individual features of each aspect may each be combined with any or all features of the other aspects. These and other aspects, features and/or technical effect will be apparent from and elucidated with reference to the illustrations described hereinafter in which:
The figures are schematic and simplified for clarity, and they just show details which are essential to the understanding of the disclosure, while other details are left out. Throughout, the same reference signs are used for identical or corresponding parts.
Further scope of applicability of the present disclosure will become apparent from the detailed description given hereinafter. However, it should be understood that the detailed description and specific examples, while indicating preferred embodiments of the disclosure, are given by way of illustration only. Other embodiments may become apparent to those skilled in the art from the following detailed description.
The detailed description set forth below in connection with the appended drawings is intended as a description of various configurations. The detailed description includes specific details for the purpose of providing a thorough understanding of various concepts. However, it will be apparent to those skilled in the art that these concepts may be practised without these specific details. Several aspects of the apparatus and methods are described by various blocks, functional units, modules, components, circuits, steps, processes, algorithms, etc. (collectively referred to as “elements”). Depending upon particular application, design constraints or other reasons, these elements may be implemented using electronic hardware, computer program, or any combination thereof.
The electronic hardware may include microprocessors, microcontrollers, digital signal processors (DSPs), field programmable gate arrays (FPGAs), programmable logic devices (PLDs), gated logic, discrete hardware circuits, and other suitable hardware configured to perform the various functionality described throughout this disclosure. Computer program shall be construed broadly to mean instructions, instruction sets, code, code segments, program code, programs, subprograms, software modules, applications, software applications, software packages, routines, subroutines, objects, executables, threads of execution, procedures, functions, etc., whether referred to as software, firmware, middleware, microcode, hardware description language, or otherwise.
The present application relates to the field of hearing devices, e.g. hearing aids.
Compared to hearing instruments (HDL, HDR) which always are positioned at essentially the same location, a body-worn microphone unit MICU, e.g. comprising a microphone array, may be positioned different each time it is mounted as illustrated by the different positions of the microphone unit MICU in
In order to achieve a good directional performance using the microphone array of the microphone unit, it may be advantageous to have a good estimate of the transfer function between the source of interest and the different microphones, or alternatively the relative (normalized) transfer functions (RTF) between the microphones with respect to the source of interest. We term this transfer function ‘the look vector d’. In other words, the look vector d is a representation of the (e.g. relative) acoustic transfer function from a target sound source (here the user's own voice, i.e. from the mouth of the user) to each microphone of the microphone unit. The look vector is preferably determined (either in advance of the use of the hearing system or adaptively) while a target (the user's voice) signal is present or dominant (e.g. present with a high probability, e.g. ≧70%) in the electric input signals of the microphones of the microphone unit. Inter-microphone covariance matrices and an eigenvector corresponding to a dominant eigenvalue of the covariance matrix are determined based thereon (cf. e.g. EP2701145A1, or EP2882204A1). The eigenvector corresponding to the dominant eigenvalue of the covariance matrix is the look vector d. The look vector depends on the relative location of the target signal to the microphones of the microphone unit (and the propagation properties of the acoustic channel from the target sound source to the respective microphones (M1, M2) of the microphone unit MICU.
There are different ways which can be used to improve the directional performance of the body-worn system.
One way is to estimate the transfer function (e.g. look vector/RTFs) from source to microphone while the person wearing the body-worn microphone unit is talking. The transfer function may e.g. be determined during phone conversations, because the person is likely to talk, if the far-away line is quiet. In an embodiment, at least one of the left and right hearing devices HDL and HDR, are configured to receive a direct electric audio signal from a telephone (representing the voice of a far-end communication partner). In an embodiment, at least one of the left and right hearing devices comprises a voice activity detector for determining whether (or with which probability) a voice is present in the received direct electric audio signal (the telephone signal).
Another way to estimate the transfer function from source to microphone is to use the hearing devices (e.g. hearing instruments) to determine when own-voice is present (presuming that the user wears hearing devices in a functional state). In an embodiment, one or both of the left and right hearing devices (HDL, HDR) comprise a voice activity detector for detecting whether a signal picked up by a microphone of the hearing device comprises a human voice. In an embodiment, the hearing device comprises a dedicated own voice detector adapted for indicating when a user's voice is present (either binary (1, 0) or with a certain probability [0, 1]). In an embodiment, the hearings device(s) and the microphone unit are adapted to establish a communication link between them allowing e.g. voice activity information to be exchanged between them.
Instead of (or in addition to) adaptively estimating the look vector d (RTFs), it is proposed to estimate a distance (D) or time delay (Δt) and possibly a direction (θ) from the microphone unit to the mouth of the user. In an embodiment, the hearing system comprises or has access to a dictionary of corresponding values of distance D (or time delay Δt) and possibly direction θ and sets of frequency dependent beamformer weights w(D, k) (or w(Δt, k)) or w(D, θ, k) (or w(Δt, θ, k)), k=1, . . . , K, where K is the number of frequency sub-bands. Each set of stored beamformer weights corresponds (in an approximation) to a range of distances (or time delays and possibly angles) around a central value. In an embodiment, the hearing system is configured to determine a current distance D′ from the user's mouth to the microphone unit and to select a set of beamformer weights w(D″, k) from the dictionary, where D″ is the distance that is closest to the current (estimated) distance D′, and to apply the set of beamformer weights to the beamformer filtering unit. This has the advantage that only one distance (or time delay, and optionally one angle) needs to be determined (a distance from the mouth to a fix point in the microphone unit, e.g. to a microphone, e.g. the closest one, e.g. M1 in
A distance between the mouth and the microphone unit (or the microphones of the microphone unit) or the propagation delay for sound between the mouth and the microphone unit (or the microphones of the microphone unit) can e.g. be determined as outlined in the following.
When the distance (D1 in
Furthermore, the microphone unit may be tilted (so that a reference direction MDREF of the microphone unit (e.g. an axis between two microphones) is not pointing in the direction of the mouth of the user).
As mentioned above, a dictionary of beamformer weights w(D, θ, k) accessible to the hearing system may allow a dynamic update of the beamformer filtering unit, purely based on the mouth to microphone unit distance and the tilt angle (without determining the look vector and inter-microphone noise covariance matrix).
If the body-worn microphone device contains a magnetic wireless link, with two or three orthogonal coils we can, based on the signal strength at each coil, determine not only the distance to the hearing instruments, but also the angle of the device with respect to the hearing instruments, and hereby also the mouth. This is illustrated in
In the embodiment of
In an embodiment, the multi-input beamformer filtering unit (of the microphone unit) comprises an MVDR filter providing filter weights wmvdr(k,m), said filter weights wmvdr(k,m) being based on a look vector d(k,m) and an inter-input unit (e.g. inter-microphone) covariance matrix Rvv(k,m) for the noise signal (the noise signal being e.g. the received signal when the user is NOT speaking), where k and m are frequency band and time frame indices, respectively.
In an embodiment, the multi-input noise reduction system is configured to adaptively estimate a current look vector d(k,m) of the beamformer filtering unit for a target signal originating from a target signal source located at a specific location relative to the user. In a preferred embodiment, the specific location relative to the user is the location of the user's mouth.
The look vector d(k,m) is an M-dimensional vector comprising elements (i=1, 2, . . . , M), the ith element di(k,m) defining an acoustic transfer function from the target signal source (at a given location relative to the input units (microphones) of the microphone unit) to the ith input unit (e.g. a microphone), or the relative acoustic transfer function from the ith input unit (microphone) to a reference input unit (microphone). The vector element di(k,m) is typically a complex number for a specific frequency (k) and time unit (m). The look vector d(k,m) may be estimated from the inter input unit covariance matrix {circumflex over (R)}ss(k,m) based on signals si(k,m), i=1, 2, . . . , M from a signal source measured at the respective input units (microphones) when the source is located at the given location (e.g. the location of the user's mouth).
The determination of the look vector d(k,m) from the inter microphone covariance matrix {circumflex over (R)}ss(k,m) and the determination of the beamformer filter weights wmvdr(k,m) from look vector d(k,m) and an inter-microphone noise covariance matrix Rvv(k,m) are e.g. described in [Kjems and Jensen; 2012].
In an embodiment, a number of sets of default beamformer weights wmvdr(Dp, k) (corresponding to a number of different distances Dp (p=1, . . . , ND) (and/or directions θq, q=1, . . . , Nθ) between the mouth and the microphones of the microphone unit) are determined in an offline calibration process, e.g. conducted in a sound studio with a head-and-torso-simulator (e.g. HATS, Head and Torso Simulator 4128C from Brad & Kjaer Sound & Vibration Measurement A/S) with play-back of voice signals from the dummy head's mouth, and a microphone unit mounted in a number of different positions on the “chest” of the dummy head (corresponding to said distances Di). In an embodiment, the sets of default beamformer weights are determined from measurements on the user (instead of the simulator). In an embodiment, the default beamformer weights are stored in a memory of the hearing system, e.g. of the microphone unit. In this way, an appropriate set of beamformer weights can be chosen and applied, when a current distance D′ (and/or angle θ′) has been determined.
In an embodiment, the beamformer weights wmvdr(k/m) are adaptively determined or selected.
The microphone unit comprises a control unit (CONT) configured to provide control of the multi-input beamformer filtering unit. The control unit (CONT) comprises a memory (MEM) storing reference values of a look vector (d) (and possibly also reference values of the noise-covariance matrices, and/or resulting beamformer weights wij). In an embodiment, a dictionary of exemplary look vectors (and/or noise-covariance matrices, and/or resulting beamformer weights w(D, θ, k)) for relevant locations of the microphone unit on the user's body, are stored in the memory (MEM). In an embodiment, the control unit (CONT) is configured to determine a current location of the microphone unit (MICU) on the user's body relative to the user's mouth. In an embodiment, the control unit is configured to select an appropriate look vector d and/or set of beamformer weights wij(D, θ, k) from the dictionary based on the currently determined location of the microphone unit. The control unit (CONT) comprises a correlation unit, e.g. a cross correlation unit, (XCOR) for determining a cross-correlation between a microphone signal INm of the hearing device (HD) (received from the hearing device via wireless link WL between the hearing device and the microphone unit (cf. dashed bold arrow in
In the embodiment, of
The wireless link (WL) between the hearing device (HD) and the microphone unit (MICU) may be based on near-field communication or radiated fields. The respective antenna and transceiver units (TU) for implementing the wireless link (WL) may comprise antenna coils as shown and discussed in connection with
In an embodiment, the signals transmitted from the hearing device to the microphone unit via the wireless link (WL) are re- or down-sampled and/or transmitted only in selected time windows to save power. This may be an allowable simplification, because the change in location or orientation of the microphone unit will generally be relatively slow. In an embodiment, the microphone signal INm of the hearing device (HD) transmitted via wireless link WL to the microphone unit is band-pass filtered to reduce the necessary bandwidth of the link (and thus power in the hearing device).
The hearing device (HD, e.g. HDL or HDR in
The hearing device further comprises a signal processing unit (SPU) for processing the resulting input signal RIN and is e.g. adapted to provide a frequency dependent gain and/or a level dependent compression and/or a transposition (with or without frequency compression) of one or more frequency ranges (bands) to one or more other frequency ranges (bands), e.g. to compensate for a hearing impairment of a user. The signal processing unit (SPU) provides a processed signal PRS. The hearing device further comprises an output unit (OU) for providing a stimulus OUT configured to be perceived by the user as an acoustic signal based on a processed electric signal PRS. In the embodiment of
The hearing system (here indicated in the hearing device (HD)) comprises a user interface (UI) allowing a user to influence functionality of the system (hearing device(s) and/or microphone unit), e.g. to enter and/or leave a mode of operation, e.g. a communication (e.g. telephone) mode. The user interface may further allow information about the current mode of operation or other information to be presented to the user, e.g. via a remote control device, such as a smartphone or other communication device with appropriate display and/or processing capabilities. Such information may include information from the microphone unit (MICU), as e.g. indicated by thin arrow in the wireless link WL from the microphone unit to the hearing device (HD) (optional microphone signals x′i and voice detecting signal VD, etc.) and further to the user interface (UI) via signal UC3.
The embodiment of a hearing system as illustrated in
In the embodiments of a hearing device (HD) in
The hearing device (HD) further comprises an output unit (OT, e.g. an output transducer) providing an enhanced output signal as stimuli perceivable by the user as sound based on the enhanced audio signal from the signal processing unit or a signal derived therefrom. Alternatively or additionally, the enhanced audio signal from the signal processing unit may be further processed and/or transmitted to another device depending on the specific application scenario.
In the embodiment of a hearing device in
The hearing device, e.g. the signal processing unit (SPU), comprises e.g. a feedback cancellation system for reducing or cancelling feedback from the output transducer (OT) to the input transducers (e.g. to IT3 and/or to the input transducers (IT1, IT2) of the BTE-part.
The hearing device (HD) exemplified in
In an embodiment, the hearing device, e.g. a hearing aid (e.g. the signal processing unit SPU), is adapted to provide a frequency dependent gain and/or a level dependent compression and/or a transposition (with or without frequency compression) of one or frequency ranges to one or more other frequency ranges, e.g. to compensate for a hearing impairment of a user.
It is intended that the structural features of the devices described above, either in the detailed description and/or in the claims, may be combined with steps of the method, when appropriately substituted by a corresponding process.
As used, the singular forms “a,” “an,” and “the” are intended to include the plural forms as well (i.e. to have the meaning “at least one”), unless expressly stated otherwise. It will be further understood that the terms “includes,” “comprises,” “including,” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. It will also be understood that when an element is referred to as being “connected” or “coupled” to another element, it can be directly connected or coupled to the other element but an intervening elements may also be present, unless expressly stated otherwise. Furthermore, “connected” or “coupled” as used herein may include wirelessly connected or coupled. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items. The steps of any disclosed method is not limited to the exact order stated herein, unless expressly stated otherwise.
It should be appreciated that reference throughout this specification to “one embodiment” or “an embodiment” or “an aspect” or features included as “may” means that a particular feature, structure or characteristic described in connection with the embodiment is included in at least one embodiment of the disclosure. Furthermore, the particular features, structures or characteristics may be combined as suitable in one or more embodiments of the disclosure. The previous description is provided to enable any person skilled in the art to practice the various aspects described herein. Various modifications to these aspects will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other aspects.
The claims are not intended to be limited to the aspects shown herein, but is to be accorded the full scope consistent with the language of the claims, wherein reference to an element in the singular is not intended to mean “one and only one” unless specifically so stated, but rather “one or more.” Unless specifically stated otherwise, the term “some” refers to one or more.
Accordingly, the scope should be judged in terms of the claims that follow.
Number | Date | Country | Kind |
---|---|---|---|
16184252.1 | Aug 2016 | EP | regional |