This invention relates, in general, to hearing tests and systems for aiding hearing and, in particular, to systems for aiding hearing, hearing aids, and methods for use of the same that provide hearing testing as well as signal processing and feature sets to enhance speech and sound intelligibility.
Traditionally, the management of hearing loss has been anchored in a process that confines the crucial step of audiogram assessment and fitting to specialized test facilities. This conventional approach necessitates that individuals seeking hearing aid adjustments must physically visit these facilities to undergo testing, followed by the fitting of the hearing aid according to the newly assessed audiogram. In the event of any changes in the patient's hearing capabilities or dissatisfaction with the hearing aid's performance, the cycle necessitates a return to the test facility for reassessment. This process not only imposes significant logistical challenges but delays the optimization of hearing aid settings to accommodate evolving patient needs. Hence, there is a burgeoning need for innovative hearing aids and methodologies that transcend these traditional constraints, offering patients the flexibility to tailor their hearing experience directly, without the repeated need to revert to test facilities for adjustments.
This application introduces a transformative approach to existing in-situ hearing aid technology, fundamentally redefining the audiogram's role and the concept of hearing testing. Unlike traditional hearing care systems, where audiograms are generated and stored at test facilities, requiring patients to visit for testing, fitting, and subsequent adjustments, this embodiment of innovation embeds the audiogram directly within the hearing aid itself and enables real-time, user-driven audiogram adjustments via a smart device, for example, effectively making the journey to test facilities for adjustments obsolete. “Testing” is reinterpreted to mean generating an up-to-the-moment audiogram through the user's smart device, allowing for immediate customization of the hearing aid's settings based on current environmental needs and personal hearing preferences. For example, a user troubled by excessive high-frequency sounds can instantly adjust the frequency settings through their smart device and upload the new audiogram directly to their hearing aid, bypassing traditional processes and devices.
Moreover, the innovation extends to the smart device conducting actual hearing tests by playing harmonics, for example, integrating seamlessly with the hearing aid to refine the audiogram. This direct integration challenges existing in-situ hearing aids and the conventional reliance on external test facilities and real-time communication between smart devices and hearing aids, which is hampered by time delays. Further, systems and methodology are presented for dissecting the frequency range into distinct frequency segments and managing each as an independent entity. This approach enhances the precision and customization of the hearing aid.
In one embodiment, a hearing aid device equipped with sound processing capabilities, including a microphone, speaker, and electronic signal processor wirelessly communicates with a smart device. The electronic signal processor is adaptable based on a custom audiogram, which can be dynamically adjusted through a smart device application and which may be stored on the hearing aid device. Patients can directly influence their hearing experience by modifying the audiogram via the app, adjusting settings like directional microphone activation, noise cancellation levels, and amplification for specific frequency ranges. This method allows users to tailor their hearing aid settings to their immediate environment and personal hearing needs, ensuring an optimized auditory experience. The system emphasizes user control and adaptability, offering a significant advancement in hearing aid technology. These and other aspects of the invention will be apparent from and elucidated with reference to the embodiments described hereinafter.
For a more complete understanding of the features and advantages of the present invention, reference is now made to the detailed description of the invention along with the accompanying figures in which corresponding numerals in the different figures refer to corresponding parts and in which:
While the making and using of various embodiments of the present invention are discussed in detail below, it should be appreciated that the present invention provides many applicable inventive concepts, which can be embodied in a wide variety of specific contexts. The specific embodiments discussed herein are merely illustrative of specific ways to make and use the invention, and do not delimit the scope of the present invention.
Referring initially to
In some embodiments, the hearing aid device 12 seamlessly integrates with a proximate smart device 14—such as a smartphone, smartwatch, tablet computer, or wearable. This integration is facilitated through a user-friendly interface displayed on the smart device 14, which hosts a range of intuitive controls including volume adjustments, operational mode selections, and real-time audiogram customization features. The user can effortlessly transmit control signals wirelessly from the smart device 14 to the hearing aid device 12, enabling immediate changes to volume, operational modes such as directional sound focus, noise cancellation levels, and audiogram adjustments, for example. This direct interaction heralds a significant shift towards user-empowered hearing aid management, allowing for on-the-spot modifications tailored to the user's specific auditory environment and personal preferences.
Central to this system is a programming interface that establishes a dynamic communication channel between the hearing aid device 12 and the smart device 14. This bidirectional interface supports the direct adjustment and real-time customization of the hearing aid's settings via an application on the smart device 14. The application empowers users to actively manage and fine-tune their hearing experience, from adjusting an audiogram 20 that may be stored on the hearing aid device 12 and displayed on the smart device 14, to selecting specific sound processing features. This level of customization and control directly from the user's smart device is unprecedented, moving beyond the traditional confines of hearing aid management and setting a new standard for personal auditory assistance. Further, this programming interface is an extensible architecture configured to integrate additional operational modes and functionalities beyond those described, wherein the architecture enables the seamless addition of features and enhancements derived from future scientific achievements, thereby allowing for continuous improvement and expansion of the system's capabilities in response to evolving user needs and technological advances.
Furthermore, this system addresses the concept of auditory testing and audiogram customization. Utilizing the smart device application, users can conduct on-demand auditory tests, thereby transforming any location into a potential test environment. This capability enables users to create and adjust their audiograms in real time, based on immediate hearing assessments and environmental conditions. This approach stands in stark contrast to conventional methods reliant on static, infrequently updated audiograms and limited adaptability. By placing the power of audiogram customization and auditory testing directly in the hands of the user, the system offers unparalleled flexibility and personalization in hearing aid technology, marking a significant leap forward from existing practices.
Referring to
As alluded, the hearing aid device 12 may be an vivo adaptare device—adapting to the living within the living—that incorporates the functionality to not only assist hearing but also to conduct auditory tests directly within the user's ear, where the hearing aid device 12 is situated. This means the hearing aid can generate, adjust, and apply audiograms-personalized hearing profiles based on the user's specific hearing capabilities and environmental conditions-without the need to remove the device or visit a professional audiologist for testing in a separate facility.
With this arrangement, the systems and methods presented herein allow the hearing aid to assess hearing capabilities in the actual environment—dynamically—where the user listens, providing more accurate and personalized results than traditional, clinic-based audiograms. Unlike traditional hearing aids, which are programmed using audiograms obtained from clinical tests, this vivo adaptare hearing aid embodiment can generate and update audiograms automatically. This process may involve playing a series of harmonic tones directly into the ear through the hearing aid and measuring the user's responses to these sounds, effectively mapping out the user's hearing profile in real-time. Once the audiogram is generated or updated, the hearing aid device 12 can immediately adjust its settings to match the user's current hearing needs. This dynamic approach allows users to have their hearing aids adjusted for optimal performance across different listening environments, such as moving from a quiet room to a noisy outdoor setting.
Referring now to
As shown, a signaling architecture communicatively interconnects the microphone inputs 138 to the electronic signal processor 130 and the electronic signal processor 130 to the speaker output 140. The various hearing aid controls 142, 144, the induction coil 146, the battery 148, and the transceiver 150 are also communicatively interconnected to the electronic signal processor 130 by the signaling architecture. The speaker output 140 sends the sound output to a speaker or speakers to project sound and in particular, acoustic signals in the audio frequency band as processed by the hearing aid 10. The hearing aid controls 142, 144 may include an ON/OFF switch as well as volume controls, for example. It should be appreciated, however, that in some embodiments, all control is manifested through the adjustment of the vivo adaptare audiogram. The induction coil 146 may receive magnetic field signals in the audio frequency band from a telephone receiver or a transmitting induction loop, for example, to provide a telecoil functionality. The induction coil 146 may also be utilized to receive remote control signals encoded on a transmitted or radiated electromagnetic carrier, with a frequency above the audio band. Various programming signals from a transmitter may also be received via the induction coil 146 or via the transceiver 150, as will be discussed. The battery 148 provides power to the hearing aid 10 and may be rechargeable or accessed through a battery compartment door (not shown), for example. The transceiver 150 may be internal, external, or a combination thereof to the housing. Further, the transceiver 150 may be a transmitter/receiver, receiver, or an antenna, for example. Communication between various smart devices and the hearing aid 10 may be enabled by a variety of wireless methodologies employed by the transceiver 150, including 802.11, 3G, 4G, Edge, WiFi, ZigBee, near field communications (NFC), Bluetooth low energy, and Bluetooth, for example.
The various controls and inputs and outputs presented above are exemplary and it should be appreciated that other types of controls may be incorporated in the hearing aid device 10. Moreover, the electronics and form of the hearing aid device 10 may vary. The hearing aid device 10 and associated electronics may include any type of headphone configuration, a behind-the-ear configuration, an in-the-ear configuration, or in-the-ear configuration, for example.
Referencing
Inside the hearing aid, the electronic signal processor 130 operates as a sophisticated computational unit, processing complex instructions stored within its memory. This memory, which can be either volatile for temporary data storage or non-volatile for long-term data retention, is crucial for the adaptive functionalities of the hearing aid. Upon execution of these instructions, the processor transforms input analog signals from the microphone into digital signals for advanced processing. This process includes an innovative step where the digital signal is adjusted based on a user-specific audiogram, incorporating a subjective assessment of sound quality. This unique feature allows the electronic signal processor 130 to personalize the audio output, tailoring it to the user's hearing profile by adjusting the signal to match the preferred hearing range identified in the audiogram. Consequently, the processor refines the digital signal into an optimized analog output, ready to be delivered to the user through the hearing aid's speaker. This end-to-end processing not only customizes sound based on individual preferences but also dynamically adapts to changing environmental conditions, significantly enhancing the auditory experience.
Further, the memory's processor-executable instructions extend the hearing aid's capabilities, enabling it to respond to various control signals for volume adjustment and mode selection. By way of example, these instructions facilitate the activation of specialized operational modes, such as directional sound focus, noise cancellation, and frequency amplification, allowing adjustments on a per-ear basis to suit different listening environments. Integration with a smart device is achieved through a wireless connection, enabled by the hearing aid's transceiver 150, which allows for real-time customization of settings via the smart device. This seamless interaction is made possible by a programming interface that supports the exchange of audiogram settings and control commands between the hearing aid and the smart device. This advanced communication empowers users to directly and effortlessly adjust their hearing aid settings, offering an unparalleled level of control and personalization. By enabling these functionalities, the hearing aid system evolves into a highly adaptable and user-centered device, capable of delivering a bespoke auditory experience that meets the unique needs of each individual.
Referring now to
In operation, the teachings presented herein permit the proximate smart device 14 such as a smart phone to form a pairing with the hearing aid device 12 and operate the hearing aid device 12. As shown, the proximate smart device 14 includes the memory 182 accessible to the processor 180 and the memory 182 includes processor-executable instructions that, when executed, cause the processor 180 to provide an interface for an operator that includes an interactive application for viewing the status of the hearing aid device 12. The processor 180 is caused to present a menu for controlling the hearing aid device 12. The processor 180 is then caused to receive an interactive instruction from the user and forward a control signal via the transceiver 186, for example, to implement the instruction at the hearing aid device 12. The processor 180 may also be caused to generate various reports about the operation of the hearing aid device 12. The processor 180 may also be caused to translate or access a translation service for the audio.
In a further embodiment of processor-executable instructions, the processor-executable instructions cause the processor 180 to create a pairing via the transceiver 186 with the hearing aid device 12. Then, the processor-executable instructions may cause the processor 400 to transform through compression with distributed computing between the processor 180 and the hearing aid device 12, the digital signal into a processed digital signal having the qualified sound range, which includes the preferred hearing range as well as the subjective assessment of sound quality, as represented by the dynamically customizable audiogram. It should be appreciated, however, that in some embodiments the distributed computing is not necessary and all functionality may be with the hearing aid device 12. The dynamically customizable audiogram may include a range or ranges of sound corresponding to highest hearing capacity of an ear of a patient modified with a subjective assessment of sound quality according to the patient. The dynamically customizable audiogram may include a completed assessment of a degree of annoyance caused to the user by an impairment of wanted sound. The dynamically customizable audiogram according to the user may also include a completed assessment of a degree of pleasantness caused to the patient by an enablement of wanted sound. That is, the subjective assessment according to the user may include a completed assessment to determine best sound quality to the user.
Significantly, the processor-executable instructions extend beyond basic device operation, enabling users to actively participate in their auditory experience. Users can select operational modes, such as directional sound mode, amplification mode, and background noise reduction mode, tailored to their immediate environmental needs and hearing preferences. This functionality is emblematic of the system's dynamic architecture, where adjustments to the hearing aid's settings are not just reactionary but predictive and personalized, fostering an auditory environment that is both adaptive and immersive.
Moreover, the integration of distributed computing between the smart device 14 and the hearing aid device 12 facilitates the transformation of digital signals into a processed digital format, reflecting the nuanced preferences captured in the dynamically customizable audiogram. That is, the processor-executable instructions receive, through the user interface, patient inputs for adjusting a dynamically customizable audiogram that represents preferred hearing settings of the patient, including adjustments to one or more frequency segments. Each of the frequency segments are a divided portion of the dynamically customizable audiogram. The processor-executable instructions then process the patient inputs to adjust the dynamically customizable audiogram, thereby enabling adaptation to varying auditory environments as perceived by the patient. The processor-executable instructions then cause the transmission of the adjusted dynamically customizable audiogram to the hearing aid device for immediate application. This dynamically customizable audiogram, adjustable via the smart device, encapsulates a spectrum of auditory capabilities and preferences, including subjective assessments of sound quality. Such assessments enable the identification and enhancement of sounds, clarity ensuring and reducing discomfort, thereby exemplifying the system's commitment to providing a tailored and enriched auditory experience for users. Through this innovative approach, the hearing aid system 10 not only adapts to the auditory landscape but also reshapes it, making it conducive to the unique needs and preferences of the user, marking a paradigm shift in personalized hearing care.
Positioned above the vivo adaptare audiogram module is the subjective assessment module 196. This module integrates the user's personal preferences and perceptions of sound quality into the hearing aid's processing algorithms. By assessing and incorporating feedback on sound clarity, volume, and tone, the subjective assessment module ensures that the audio output is finely tuned to the user's specific auditory preferences, enhancing the overall satisfaction with the hearing aid's performance. Further enhancing the system's functionality are several function modules, labeled 198, 200, and 202, each designed to perform distinct sound processing tasks. Function module 198 serves as an advanced equalizer, offering precise control over frequency response to shape the audio signal according to the user's customized audiogram and subjective preferences. This allows for the attenuation or amplification of specific frequency bands, ensuring that the sound delivered to the user is both clear and comfortable. Adjacent to the equalizer, additional function modules (200) provide various specialized processing capabilities, such as noise reduction, feedback suppression, frequency transition, dead zone analysis, and spatial awareness, further refining the sound and quality intelligibility for the user. The series culminates with function module 202, acting as a sophisticated amplifier. This module is responsible for adjusting the overall volume of the audio signal to the optimal listening level as determined by the vivo adaptare audiogram, subjective assessments, and user-controlled settings. The amplifier ensures that the sound is delivered at a consistent, comfortable level, accommodating for both the quiet and loud acoustic environments the user may encounter. Together, these modules within system 10 depict a holistic approach to hearing aid design, emphasizing personalization, adaptability, and user control. By integrating these advanced modules, the system offers a tailored auditory experience, significantly surpassing the capabilities of traditional hearing aids.
Delving into
Transitioning to
In an illustrative leap to
Turning attention to
Expanding upon the customization options, Controls 262 introduce the user to the capability of setting, storing, and uploading changes to a series of dynamically customizable audiograms, labeled as 270, 272, and 274. These audiograms represent specific auditory profiles tailored to different listening environments and preferences. The dynamically customizable audiogram 270 illustrates an adjustment in volume, allowing users to modify the overall loudness of the hearing aid output to achieve the desired auditory balance. The dynamically customizable audiogram 272 is dedicated to background noise suppression, enabling users to create a hearing profile that effectively reduces ambient noise, thus enhancing the clarity and intelligibility of foreground sounds. Lastly, the dynamically customizable audiogram 274 focuses on compression adjustments, providing users the ability to set a hearing profile that optimizes the dynamic range of sounds, ensuring that all sounds, regardless of their original volume, are heard comfortably and clearly.
These interfaces and controls underscore the smart device's commitment to delivering a highly personalized and adaptable hearing aid experience. By empowering users with the ability to precisely adjust and save multiple audiogram settings, the system acknowledges the dynamic nature of human hearing and the diverse auditory environments encountered in daily life. This approach not only enhances the user's autonomy over their hearing experience but also fosters a sense of engagement and satisfaction with the hearing aid system, marking a significant advancement in the integration of technology and personalized care in the realm of auditory assistance.
Referring now to
Referring now to
Proceeding to block 324, the smart device processes the patient inputs. This step involves analyzing the adjustments to ensure they align with the user's hearing preferences and the acoustic characteristics of the current environment. At block 326, the audiogram within the hearing aid is dynamically updated to reflect the processed adjustments. This crucial step enables the hearing aid to adapt its audio processing in real-time to suit the varying auditory environments experienced by the user, ensuring an optimal listening experience under diverse conditions. Finally, at block 328, the updated, customized audiogram is utilized.
The order of execution or performance of the methods and data flows illustrated and described herein is not essential, unless otherwise specified. That is, elements of the methods and data flows may be performed in any order, unless otherwise specified, and that the methods may include more or less elements than those disclosed herein. For example, it is contemplated that executing or performing a particular element before, contemporaneously with, or after another element are all possible sequences of execution.
While this invention has been described with reference to illustrative embodiments, this description is not intended to be construed in a limiting sense. Various modifications and combinations of the illustrative embodiments as well as other embodiments of the invention, will be apparent to persons skilled in the art upon reference to the description. It is, therefore, intended that the appended claims encompass any such modifications or embodiments.
This application claims priority from U.S. Provisional Patent Application Ser. No. 63/564,110 entitled “System for Aiding Hearing and Method for Use of Same” filed on Mar. 12, 2024 in the name of Laslo Olah, and U.S. Provisional patent application Ser. No. 63/632,371 entitled “System for Aiding Hearing and Method for Use of Same” filed on Apr. 10, 2024 in the name of Laslo Olah; both of which are hereby incorporated by reference, in entirety, for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
5987147 | Nishimoto | Nov 1999 | A |
7113589 | Mitchler | Sep 2006 | B2 |
8565460 | Takagi et al. | Oct 2013 | B2 |
8761421 | Apfel | Jun 2014 | B2 |
9232322 | Fang | Jan 2016 | B2 |
9344814 | Rasmussen | May 2016 | B2 |
9712928 | Pedersen et al. | Jul 2017 | B2 |
10181328 | Jensen et al. | Jan 2019 | B2 |
20050004691 | Edwards | Jan 2005 | A1 |
20050245221 | Beyer | Nov 2005 | A1 |
20080004691 | Weber | Jan 2008 | A1 |
20120106762 | Kornagel | May 2012 | A1 |
20120121102 | Jang | May 2012 | A1 |
20130142369 | Zhang et al. | Jun 2013 | A1 |
20130223661 | Uzuanis | Aug 2013 | A1 |
20160338622 | Chen | Nov 2016 | A1 |
20180035216 | Van Hasselt et al. | Feb 2018 | A1 |
20180207167 | Dyhrfjeld-Johnsen | Jul 2018 | A1 |
20200268260 | Tran | Aug 2020 | A1 |
20210006909 | Olah et al. | Jan 2021 | A1 |
Number | Date | Country |
---|---|---|
20170026786 | Mar 2017 | KR |
2016188270 | Dec 2016 | WO |
2019136382 | Jul 2019 | WO |
Entry |
---|
International Preliminary Report on Patentability dated Jul. 7, 2020 regarding International Application No. PCT/US2019/012550; 8 pp. |
International Search Report and Written Opinion dated Mar. 22, 2019 regarding International Application No. PCT/US2019/012550; 10 pp. |
International Search Report and Written Opinion dated Aug. 5, 2021 concerning International Application No. PCT/US2021/029414; 8 pp. |
Number | Date | Country | |
---|---|---|---|
63632371 | Apr 2024 | US | |
63564110 | Mar 2024 | US |