This application is related to, and incorporates by reference, U.S. patent application Ser. No. 15/106,989, filed Jun. 21, 2016 (abandoned); application Ser. No. 15/348,400, filed Nov. 10, 2016 (now U.S. Pat. No. 9,906,636); and application Ser. No. 15/352,034, filed Nov. 17, 2016 (now U.S. Pat. No. 9,799,188), all titled Intelligent Earplug System. It is also related to U.S. patent application Ser. No. 15/267,567 (now U.S. Pat. No. 10,434,279), entitled Sleep Assistance Device; application Ser. No. 15/267,464 (now U.S. Pat. No. 10,517,527), entitled Sleep Quality Scoring and Improvement; application Ser. No. 15/267,552, entitled Intelligent Wake-Up System; application Ser. No. 15/267,848 (now U.S. Pat. No. 10,653,856), entitled Sleep System; application Ser. No. 15/267,858 (now U.S. Pat. No. 10,963,146), entitled User Interface fora Sleep System; and application Ser. No. 15/267,886 (U.S. Pat. No. 10,561,362), entitled Sleep Assessment Using a Home Sleep System, all of which were filed on Sep. 16, 2016. It is also related to U.S. patent application Ser. No. 15/655,836 (now U.S. Pat. No. 10,478,590), titled Sleep Assistance Device For Multiple Users, and Ser. No. 15/655,845 (now U.S. Pat. No. 10,682,491), Ser. No. 15/655,853 (now U.S. Pat. No. 10,632,278), and Ser. No. 15/655,857 (now U.S. Pat. No. 10,848,848), all titled Earphones for Measuring and Entraining Respiration, filed Jul. 20, 2017, and incorporated here by reference.
This disclosure relates to earphones for measuring and entraining respiration.
Sleeplessness and poor or interrupted sleep may significantly affect a person's health. Poor sleep may be caused by such factors as ambient noise, stress, medical conditions, or discomfort. Thus, there exists a need for a sleep aid that can help address the underlying causes of poor sleep without adversely affecting the user's health in other, unintended ways.
In general, in one aspect, a system includes an earphone with a loudspeaker, an accelerometer, a housing supporting the loudspeaker and accelerometer, and an ear tip surrounding the housing and configured to acoustically couple the loudspeaker to an ear canal of a user, and a processor. The processor receives input signals from the accelerometer in response to motion of the earphone, and detects peaks in the input signals. Based on the detected peaks, the processor computes an instantaneous heart rate, measures a frequency of an oscillation within the instantaneous heart rate, and based on the frequency of the oscillation, computes a rate of respiration.
Implementations may include one or more of the following, in any combination. The accelerometer may be located on a circuit board within the earphone. The housing may include an outer shell of the earphone, and the accelerometer may be located in the outer shell. The input signals may include acceleration measurements for each of three axes, and the processor may detect the peaks by combining the three acceleration measurements into the norm of the three measurements. The processor may receive second input signals from a second accelerometer in a second earphone, including second acceleration measurements for each of three axes of the second accelerometer, combine the second three acceleration measurements into the norm of the second three measurements, detect the peaks by combining the norm of the three measurements from the first accelerometer with the norm of the three measurements from the second accelerometer, and detect peaks within the result of the combination. Combining the norms may include multiplying the amplitude of the norm of the first three acceleration measurements by the amplitude of the norm of the second three acceleration measurements, at each time that the two accelerometers' measurements are sampled. The processor may adjust output audio signals based on the rate of respiration, and provide the output audio signals to the loudspeaker.
Adjusting the output audio signals may include adjusting a rhythm of the output audio signals to be about one cycle per minute less than the detected respiration rate. Adjusting the output audio signals may include transitioning the output audio signals from respiration entrainment sounds to masking sounds. Adjusting the output audio signals may include transitioning the output audio signals from masking sounds to awakening sounds. The earphone may include a memory storing sound files, and providing the output audio signals may include retrieving a first sound file from the memory. Adjusting the output audio signals may include retrieving a second sound file from the memory and using the second sound file to generate the output audio signal.
The processor may be integrated within the earphone. The processor may be integrated within a portable computing device. The processor may measure the frequency of the oscillation within the instantaneous heart rate by computing a fast Fourier transform (FFT) of the instantaneous heart rate. The processor may measure the frequency of the oscillation within the instantaneous heart rate by computing a gradient of the instantaneous heart rate, and computing a fast Fourier transform (FFT) of the gradient of the instantaneous heart rate. The processor may measure the frequency of the oscillation within the instantaneous heart rate by detecting peaks of the instantaneous heart rate. The processor may measure the frequency of the oscillation within the instantaneous heart rate by fitting a sine function to the instantaneous heart rate, the frequency of the sine being the frequency of the oscillation. The ear tip may be configured to acoustically close the entrance to the user's ear canal.
In general, in one aspect, measuring the respiration rate of a user of an earphone includes receiving input signals from an accelerometer supported within the earphone, the input signals including acceleration measurements for each of three axes, and, in a processor, combining the three acceleration measurements into the norm of the three measurements, detecting peaks in the norm of the accelerometer measurements, based on the detected peaks, computing an instantaneous heart rate, measuring a frequency of an oscillation within the instantaneous heart rate, and based on the frequency of the oscillation, computing a rate of respiration.
Implementations may include one or more of the following, in any combination. Measuring the frequency of the oscillation within the instantaneous heart rate may include computing a fast Fourier transform (FFT) of the instantaneous heart rate. Measuring the frequency of the oscillation within the instantaneous heart rate may include computing a gradient of the instantaneous heart rate, and computing a fast Fourier transform (FFT) of the gradient of the instantaneous heart rate. Measuring the frequency of the oscillation within the instantaneous heart rate may include detecting peaks of the instantaneous heart rate. Measuring the frequency of the oscillation within the instantaneous heart rate may include fitting a sine function to the instantaneous heart rate, the frequency of the sine being the frequency of the oscillation. Second input acceleration signals may be received from a second accelerometer supported within a second earphone, the second input signals including acceleration measurements for each of three axes of the second accelerometer; the processor may detect the peaks by combining the second three acceleration measurements into the norm of the second three measurements, combining the norm of the first three acceleration measurements with the norm of the second three acceleration measurements, and detecting peaks within the result of the combination.
Advantages include sensing the respiration rate at the ear without interference from audio signals being generated by the earphone, and without a need to seal the ear interface.
All examples and features mentioned above can be combined in any technically possible way. Other features and advantages will be apparent from the description and the claims.
Several of the above-referenced applications describe a bedside system that detects a user's respiration rate and uses that to infer and manage their sleep state. In particular, to assist the user with falling to sleep, the system plays sounds that have a rhythm slightly slower than the user's own respiration rate. This naturally leads the user to slow their breathing to match the rhythm of the sounds, in a process referred to as entrainment. As the user slows their rate of respiration, the rate of the sounds is further reduced, in a feedback loop that leads the user gradually to sleep. Once the user falls asleep (as indicated by artifacts in their respiration rate), the system switches to playing masking sounds, which diminish the user's ability to detect, and be disturbed by, external sounds. If the user is detected to be waking up too early, entrainment may be reactivated. When it is time for the user to wake up, the system may coordinate wake-up sounds with the user's sleep state and other information to wake the user in the least-disruptive way possible.
Others of the above-referenced applications describe intelligent earplugs which the user can wear while sleeping, and which provide masking sounds through the night, and alarm or alert sounds when needed. These earplugs are controlled by a smartphone, but principally operate autonomously, playing stored masking sounds until instructed otherwise by the controlling phone, or based on an internal clock. It would be advantageous if the intelligent earplugs could play the respiration-entraining sounds of the bedside systems, to help the user fall asleep without disturbing others who may be sharing the bed or room. One solution to that, described in co-pending application Ser. No. 15/655,836, is for the sleep system to inform the earplugs of the user's respiration rate and sleep state, and for the earplugs to adjust the rate of a rhythmic component in stored entrainment sounds as in the out-loud system.
The Ser. Nos. 15/655,853 and 15/655,857 applications describe how to add respiration sensing to the earplugs themselves, using microphones, so that the external system is not required, and the earplugs can operate fully autonomously, or with only a smart phone to control them.
The technique described in the '853 application, of using a microphone to measure heart rate variability, and from that deriving respiration rate, can also be performed using an accelerometer, as described below.
As shown in
In the example of
The heartbeat is derived from the accelerometer signals, and the respiration rate is derived from the heart rate variability (HRV).
Instead of separately using the axes, we can combine the three axes of each ear into their norm:
N=√{square root over ({umlaut over (x)}2+ÿ2+{umlaut over (z)}2)} (1)
Graphs of the two norms, one for each ear, are shown in
Combining the amplitudes of the two signals at each time sample, such as by multiplication, can greatly increase the signal to noise ratio, as shown in
Applying a peak-finding algorithm to the combined accelerometer signals and observing the distance between consecutive peaks yields the beat-to-beat, or instantaneous, heart rate value, shown in
Embodiments of the systems and methods described above comprise computer components and computer-implemented steps that will be apparent to those skilled in the art. For example, it should be understood by one of skill in the art that the computer-implemented steps may be stored as computer-executable instructions on a computer-readable medium such as, for example, hard disks, optical disks, solid-state disks, flash ROMS, nonvolatile ROM, and RAM. Furthermore, it should be understood by one of skill in the art that the computer-executable instructions may be executed on a variety of processors such as, for example, microprocessors, digital signal processors, and gate arrays. For ease of exposition, not every step or element of the systems and methods described above is described herein as part of a computer system, but those skilled in the art will recognize that each step or element may have a corresponding computer system or software component. Such computer system and software components are therefore enabled by describing their corresponding steps or elements (that is, their functionality), and are within the scope of the disclosure.
A number of implementations have been described. Nevertheless, it will be understood that additional modifications may be made without departing from the scope of the inventive concepts described herein, and, accordingly, other embodiments are within the scope of the following claims.
Number | Name | Date | Kind |
---|---|---|---|
5143078 | Mather et al. | Sep 1992 | A |
5238001 | Gallant et al. | Aug 1993 | A |
7850619 | Gavish et al. | Dec 2010 | B2 |
8199956 | Haartsen et al. | Jun 2012 | B2 |
9635452 | Cheng et al. | Apr 2017 | B2 |
9690376 | Davis et al. | Jun 2017 | B2 |
20020091049 | Hisano et al. | Jul 2002 | A1 |
20040077934 | Massad | Apr 2004 | A1 |
20040111039 | Minamiura et al. | Jun 2004 | A1 |
20040243005 | Rapps | Dec 2004 | A1 |
20070113649 | Bharti et al. | May 2007 | A1 |
20070118011 | Harrison et al. | May 2007 | A1 |
20080107287 | Beard | May 2008 | A1 |
20080146890 | LeBoeuf et al. | Jun 2008 | A1 |
20080171945 | Dotter | Jul 2008 | A1 |
20090143636 | Mullen et al. | Jun 2009 | A1 |
20100125218 | Haartsen et al. | May 2010 | A1 |
20100240945 | Bikko | Sep 2010 | A1 |
20110295083 | Doelling et al. | Dec 2011 | A1 |
20110301477 | Hughes | Dec 2011 | A1 |
20120156933 | Kreger et al. | Jun 2012 | A1 |
20120203077 | He et al. | Aug 2012 | A1 |
20120225412 | Wagner | Sep 2012 | A1 |
20120256933 | Airey et al. | Oct 2012 | A1 |
20120296219 | Chon et al. | Nov 2012 | A1 |
20130034258 | Lin | Feb 2013 | A1 |
20130090567 | Lee et al. | Apr 2013 | A1 |
20130261771 | Kate | Oct 2013 | A1 |
20130272530 | Gauger, Jr. | Oct 2013 | A1 |
20140141395 | Gavish et al. | May 2014 | A1 |
20140371632 | Al-Ali | Dec 2014 | A1 |
20150100141 | Hughes | Apr 2015 | A1 |
20150125832 | Tran | May 2015 | A1 |
20150150515 | Strachan | Jun 2015 | A1 |
20150154950 | Ring | Jun 2015 | A1 |
20150164380 | O'Dwyer et al. | Jun 2015 | A1 |
20150258301 | Trivedi et al. | Sep 2015 | A1 |
20150313552 | Zhang et al. | Nov 2015 | A1 |
20150351688 | Just et al. | Dec 2015 | A1 |
20160151603 | Shouldice et al. | Jun 2016 | A1 |
20160331273 | Armoundas | Nov 2016 | A1 |
20170061760 | Lee et al. | Mar 2017 | A1 |
20170094385 | Lee et al. | Mar 2017 | A1 |
20170149945 | Lee et al. | May 2017 | A1 |
20170195811 | Yen et al. | Jul 2017 | A1 |
20170258329 | Marsh | Sep 2017 | A1 |
20170325718 | Boesen et al. | Nov 2017 | A1 |
20170347177 | Masaki et al. | Nov 2017 | A1 |
20170367658 | LeBoeuf et al. | Dec 2017 | A1 |
20180014741 | Chou | Jan 2018 | A1 |
20180035916 | Hirose | Feb 2018 | A1 |
20180078197 | Ware et al. | Mar 2018 | A1 |
20180078198 | Reich et al. | Mar 2018 | A1 |
20180078732 | Keshavan et al. | Mar 2018 | A1 |
20180078733 | Freed et al. | Mar 2018 | A1 |
20180081527 | Dolecki et al. | Mar 2018 | A1 |
20180082550 | Read et al. | Mar 2018 | A1 |
20180338193 | Wallace et al. | Nov 2018 | A1 |
20190022348 | Read et al. | Jan 2019 | A1 |
20190022349 | Kirszenblat et al. | Jan 2019 | A1 |
20190028789 | Stockton, X et al. | Jan 2019 | A1 |
20190231198 | Hirota et al. | Aug 2019 | A1 |
20190239772 | Grace et al. | Aug 2019 | A1 |
Number | Date | Country |
---|---|---|
1003244 | Apr 2012 | BR |
1550204 | Dec 2004 | CN |
1502298 | Jul 2007 | CN |
102104815 | Jun 2011 | CN |
102215740 | Oct 2011 | CN |
103750831 | Apr 2014 | CN |
102293013 | Sep 2014 | CN |
203841684 | Sep 2014 | CN |
104207755 | Dec 2014 | CN |
104257354 | Jan 2015 | CN |
104434064 | Mar 2015 | CN |
104737554 | Jun 2015 | CN |
104822320 | Aug 2015 | CN |
105027030 | Nov 2015 | CN |
105430546 | Mar 2016 | CN |
106072962 | Nov 2016 | CN |
106725488 | May 2017 | CN |
106851459 | Jun 2017 | CN |
106999048 | Aug 2017 | CN |
107095659 | Aug 2017 | CN |
107198516 | Sep 2017 | CN |
107405106 | Nov 2017 | CN |
103270522 | Jan 2018 | CN |
69204555 | May 1996 | DE |
10014077 | Oct 2001 | DE |
102006057709 | Jun 2008 | DE |
0640262 | Mar 1995 | EP |
2392256 | Dec 2011 | EP |
2729058 | May 2014 | EP |
2758709 | Feb 1999 | FR |
2532745 | Jun 2016 | GB |
09248282 | Sep 1997 | JP |
2003339652 | Dec 2003 | JP |
2008241717 | Oct 2008 | JP |
WO-9209232 | Jun 1992 | WO |
2010054863 | May 2010 | WO |
2014101913 | Jul 2014 | WO |
WO-2015013450 | Jan 2015 | WO |
2016181148 | Nov 2016 | WO |
2017011431 | Jan 2017 | WO |
2017068571 | Apr 2017 | WO |
2017203251 | Nov 2017 | WO |
Entry |
---|
Axel Schäfer et al: “Estimation of Breathing Rate from Respiratory Sinus Arrhythmia: Comparison of Various Methods”, Annals of Biomedical Engnieering, Kluwer Academic Publishers—Plenum Publishers, NE, vol. 36, No. 3, Jan. 11, 2008 (Jan. 11, 2008), pp. 476-485, XP019568785, ISSN: 1573-9686, p. 476-p. 480. |
CN Office Action dated Jan. 6, 2022 from CN Application 201880048094.6. |
D.Oletic and V. Bilas, “Energy-Efficient Respiratory Sounds for Personal Mobile Asthma Monitoring”, in IEEE Sensors Journal, vol. 16, No. 23, pp. 8295-8303, Dec. 1, 2016. |
International Search Report and Written Opinion dated Oct. 10, 2018 for PCT/US2018/043058. |
International Search Report and Written Opinion dated Oct. 10, 2018 for PCT/US2018/043061. |
International Search Report and Written Opinion dated Oct. 10, 2018 for PCT/US2018/043062. |
International Search Report and Written Opinion dated Apr. 26, 2019 for PCT/US2019/014772. |
International Search Report and Written Opinion dated Jul. 3, 2019 for PCT/US2019/014066. |
Invitation to Pay Additional Fees and Partial International Search Report dated May 2, 2019 for PCT/US2019/014066. |
Kim J M et al: “Two Algorithms for Detecting Respiratory Rate from ECG Signal”, Imaging the Future Medicine : World Congress on Medical Physics and Biomedical Engineering, Aug. 27, 2006-Sep. I, 2006, Coex Seoul, K; [IFMBE Proceedings, vol. 14], Springer, DE, vol. 14, Jan. 1, 2007 (Jan. 1, 2007), pp. 4069-4071, XP008169469, DOI: 10 .1007/978-3-540-36841-0 1030 ISBN : 978-3-540-36839-7—the whole document. |
Richard Singhathip et al: “Extracting Respiration Rate From Raw ECG Signals”, Biomedical Engineering: Applications, Basis and Communications = Yixue-Gongc, World Scientific, TW, vol. 22, No. 4, Aug. 1, 2010 (Aug. 1, 2010), pp. 307-314, XP008169483, ISSN : 1016-2372, DOI: 10.4015/S1016237210002079 pp. 307-p. 309. |
First Office Action from CN App. No. 201980012599.1 dated Mar. 11, 2022, 12 pages. |
Number | Date | Country | |
---|---|---|---|
20210251494 A1 | Aug 2021 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15881034 | Jan 2018 | US |
Child | 17308438 | US |