Hardware sensor system for improved sleep detection

Information

  • Patent Grant
  • 12048529
  • Patent Number
    12,048,529
  • Date Filed
    Monday, October 17, 2022
    2 years ago
  • Date Issued
    Tuesday, July 30, 2024
    4 months ago
Abstract
A hardware sensor system comprising a piezo sensor outputting charge data corresponding to a motion, an insulated cable from the piezo sensor to a receiver, to transmit the charge data, an insulated charge to voltage converter on the receiver, the insulated charge to voltage converter converting the charge data to voltage data, an analog-to-digital converter to convert the voltage data to digital data, and an uploader to upload the data to a server for processing.
Description
FIELD

The present invention relates to sleep sensors, and more particularly to an improved sleep detection system including sensor hardware.


BACKGROUND

An average person spends about one-third of his or her life asleep. Sleep is the time our bodies undergo repair and detoxification. Research has shown that poor sleep patterns is an indication of and often directly correlated to poor health. Proper, restful and effective sleep has a profound effect on our mental, emotional and physical well-being.


Every person has a unique circadian rhythm that, without manipulation, will cause the person to consistently go to sleep around a certain time and wake up around a certain time. For most people, a typical night's sleep is comprised of five different sleep cycles, each lasting about 90 minutes. The first four stages of each cycle are often regarded as quiet sleep or non-rapid eye movement (NREM). The final stage is often denoted by and referred to as rapid eye movement (REM). REM sleep is thought to help consolidate memory and emotion. REM sleep is also the time when blood flow rises sharply in several areas of the brain that are linked to processing memories and emotional experiences. During REM sleep, areas of the brain associated with complex reasoning and language experience blood flow declines, whereas areas of the brain associated with processing memories and emotional experiences exhibit increased blood flow.


Therefore, it is useful for everyone to know more about how well they sleep.





BRIEF DESCRIPTION OF THE FIGURES

The present invention is illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:



FIG. 1 is block diagram of one embodiment of a system in which the sensor system may be implemented.



FIG. 2 is a block diagram of one embodiment of the sensor and receiver portion of the sensor system.



FIG. 3 is a block diagram of one embodiment of the first portion of the receiver and the A-to-D converter.



FIG. 4 is a circuit diagram of one embodiment of the first portion of the receiver.



FIG. 5 is a circuit diagram of one embodiment of the A-to-D converter.



FIG. 6 is a block diagram of one embodiment of the digital portion of the receiver.



FIG. 7 is a flowchart of one embodiment of data collection through upload to the server.





DETAILED DESCRIPTION

A sleep monitoring system is described. The system includes analog and digital elements, which collect data from a sleeper and provide it for processing and analysis to a server system. In one embodiment, the sleep monitoring system includes a sensor which is designed to be placed under a user's mattress or mattress topper, or in a user's bedframe. In one embodiment, this sensor collects movement data, and sends it through an insulated cable to a receiver. In another embodiment, the data may be sent wirelessly. The receiver, which in one embodiment is positioned in proximity to the bed, receives the insulated cable or wireless signal, and converts the data to a digital signal. In one embodiment, the digital signal is uploaded to the server for further processing and analysis. The server analyzes the sleep data, and can be used to set the receiver's operation, as well as control the user's sleep environment, in one embodiment.


The following detailed description of embodiments of the invention makes reference to the accompanying drawings in which like references indicate similar elements, showing by way of illustration specific embodiments of practicing the invention. Description of these embodiments is in sufficient detail to enable those skilled in the art to practice the invention. One skilled in the art understands that other embodiments may be utilized, and that logical, mechanical, electrical, functional and other changes may be made without departing from the scope of the present invention. The following detailed description is, therefore, not to be taken in a limiting sense, and the scope of the present invention is defined only by the appended claims.



FIG. 1 is block diagram of one embodiment of a system in which the sensor system may be implemented. The system includes a sleep analytics system 100 including sensors 120, receiver 130, server 160, and analytics engine 170. In one embodiment, the client portion of the sleep analytics system 100 is located in a user's home includes the sensors 120 and receiver 130.


In one embodiment, the receiver 130 is coupled to sensors 120 via a cable. In another embodiment the connection may be wireless, such as low power Bluetooth (BLE), Wi-Fi, or another type of wireless connection. In one embodiment, receiver 130 also may be coupled to a controller 140, which controls bed 150. In one embodiment, this connection is a wired connection. Alternatively, it may be a wireless connection.


In one embodiment, the sensors 120 may include one or more sensors positioned in bed 150 which are used to measure the user's sleep. In one embodiment, sensors 120 may include sensors which are not in bed 150 but positioned in the room in which the bed 150 is located. In one embodiment, one or more these additional sensors may be built into receiver 130. In one embodiment, there may be external sensors which may be coupled to receiver 130 either via wires or wirelessly. The receiver 130 collects data from the one or more sensors, for transmission to the server 160.


In one embodiment, the receiver 130 is coupled to the server 160 via a network 150. The server portion includes server 160 and analytics engine 170, which in one embodiment are located off-site, removed from the user. In another embodiment, the server may be a local system, such as a computer system running an application. The network 150 may be the Internet, and the receiver 130 may send data to the server via a wireless network, such as Wi-Fi or the cellular network. In one embodiment, server 160 and analytics engine 170 may be on the same physical device. In one embodiment, server and/or analytics engine 170 may include a plurality of devices. In one embodiment, one or both of the server 170 and the analytics engine 170 may be using cloud computing and may be implemented as a distributed system.



FIG. 2 is a block diagram of one embodiment of the sensor and receiver portion of the sensor system. In one embodiment, sleep sensors 210 include two sensors 212, 214, which are designed to be placed underneath a mattress, mattress topper, or mattress cover. In one embodiment, the sensors are piezoelectric sensors positioned on a hard foam surface to provide compressibility and support. In one embodiment, the sensors are coupled to an audio codec 225 to encode the data, for transmission to the server. By using audio codec, the data can be encoded in a way that provides 100% accuracy and a reduction in data size. In one embodiment, the receiver further includes additional environmental sensors 225. In one embodiment, the environmental sensors 225 may include a humidity and temperature sensor and a volatile organic compounds (VOC) sensor. Other sensors may also be included in the receiver.


In one embodiment, he encoded data from the sleep sensors 210 and the data from the environmental sensors 225 (which may also be encoded in some embodiments) are passed to a processor and Wi-Fi module 230. The processor and Wi-Fi module 230 sends the data to the server via a Wi-Fi connection 270. In another embodiment, the processor and Wi-Fi module 230 may be replaced by a separate processor and network access element. The Wi-Fi module may be replaced by a mobile network chip. In one embodiment, the processor and Wi-Fi module 230 includes a random access memory, such as DDR2, to buffer the data from the sensors, prior to transmission. In one embodiment, a flash memory may store the code for the processor 230.


In one embodiment, power subsystem 250 provides power to the processor 230, codec 225, and environmental sensors. In one embodiment, the power subsystem 250 provides a 3.3V power to the processor 230. In one embodiment, the power subsystem 250 also provides 1.8V to the audio codec 225, or other elements of the system. In one embodiment, the sleep sensors 210 provide their data as charge data between 2.5V and −2.5V. Thus, in one embodiment, the power subsystem handles voltages between −2.5V and 5V.



FIG. 3 is a block diagram of one embodiment of the first portion of the receiver and the A-to-D converter. In one embodiment, the system includes an insulated sensor 310. The insulated sensor is a piezo sensor, which is sensitive to movement. The sensor 310 itself is insulated, to ensure that it is not impacted by stray signals. Because the sensor 310 provides data as a voltage level, any noise impacting the sensor 310 may overwhelm the real data. In one embodiment, a ground connection is provided to the sensor 310 to provide insulation. The sensor 310 is coupled to the receiver 300 via cable 320, in one embodiment. In one embodiment, the cable 320 is a grounded cable. In one embodiment, the cable 320 is a custom shielded cable with two conductors.


The cable 320 connects to a plug 330 in receiver. In one embodiment, the plug is an insulated plug, to shield the data from the sensor from noise. In one embodiment, an electrostatic discharge protector (EDS) 340 is coupled to the line as well. A charge to voltage converter 350 converts the output of the sensors 310, which is charge data, into a voltage. The plug 330, EDS protector 340, and charge to voltage converter 350 all are on an insulated portion 220 of the receiver. In one embodiment, a custom metal enclosure provides the insulation. In one embodiment, the custom metal enclosure is grounded. In one embodiment, the receiver utilizes a three-prong plug, to request a grounded outlet. In one embodiment, the receiver verifies that the wall connection provides a proper ground. In one embodiment, the user may be alerted if the receiver is plugged into an ungrounded outlet. However, in one embodiment, the metal enclosure provides protection/insulation even when not properly grounded.


The output of the charge to voltage converter 350 is passed to an input conditioner, in one embodiment. The input conditioner 360 adjusts the voltage range of the signal. The voltage is then passed to an analog to digital converter 370, in one embodiment. This converts the analog sensor data into a digital signal. The output of the analog to digital converter 370 is encoded and sent to the server for analysis. In one embodiment, the digital signal is encoded to ensure error correction. The signal may also be compressed.


For simplicity this figure, and others, illustrate a single sensor and connection. In one embodiment, the system may include two sensors when configured to detect two sleepers. In one embodiment, the system may include more than two sensors. When additional sensors are used, they may be separately handled. In one embodiment, each sensor has a separate and substantially identical path. In another embodiment, multiple sensors may send their data to the receive through a shared path.



FIG. 4 is a circuit diagram of one embodiment of the first portion of the receiver. The receiver includes the sensor connector 410 to which the cable is coupled. ESD protection diode 420 is tied to ground and protects against electrostatic discharge.


The charge mode amplifier 430 provides a charge to voltage conversion for the signal from the sensors. The charge mode amplifier is an op-amp with a negative feedback capacitor and a large resistor 435, converting the charge signal to a voltage output.


The voltage output from the charge mode amplifier 430 is passed to an input conditioning amplifier 440. The input conditioning amplifier is 440 an op-amp that adjusts the voltage range of the signal, for the analog-to-digital converter. In one embodiment, the input to the input conditioning amplifier a voltage value between −2.5V and 2.5V and adjusts it to 0V to 1.8V. In one embodiment, this element may be skipped if the Analog-to-Digital converter can handle the voltage range output by the charge mode amplifier 430.



FIG. 5 is a circuit diagram of one embodiment of the A-to-D converter. In one embodiment, for simplicity and to ensure that the two signals are processed in a synchronized manner, the A-to-D converter 510 is an Audio CODEC, which provides concurrent sampling of two channels, at 24 bits. This maintains time alignment between data from the two sensors. Of course, another type of analog-to-digital converter may be used. In one embodiment, the A-to-D converter used should provide at least two channels, and at least an 18-bit rate sampling. This is the last portion of the receive which is analog. The output of the A-to-D converter is digital and is passed to the digital portion of the receiver. In one embodiment, the analog portion of the receiver and the digital portion are on separate substrates.



FIG. 6 is a block diagram of one embodiment of the digital portion of the receiver. The data from the A-to-D converter in one embodiment is 24-bit two channel data.


This is input into a downsampler 610, in one embodiment. In one embodiment, the receiver selects which downsampler to utilize. In one embodiment, the server controls the receiver's selection. In one embodiment, the selection is based on the data received and analyzed by the server.


The reason for sample rate selection is to optimize the upload based on a current state of the sleep monitor. In one embodiment, the sleep monitor states may include: not in use (no one on bed), in use for limited analysis, and in full use. For example, in one embodiment, if no one is on the bed, the rate can be downsampled to a lowest sample rate, for example between 0.5 and 5 Hz. In one embodiment, the lowest sample rate is 1-Hz. If only sleep-states and HR/BR measurements are being utilized, the sample rate can be reduced to a mid-range frequency, for example 30 to 100 Hz. In one embodiment, the midrange frequency is 80 Hz. 80-Hz. Whereas, if sleep states, HR/BR, snore detection, respiratory events, HRV, etc., are being measured, a higher rate, for example 100 Hz to 500 Hz may be used. In one embodiment, the higher rate is a 320-Hz rate. In one embodiment, the sample mode selector 620 determines the sampling rate. In one embodiment, software services running on the cloud determine and remotely set the sample rate selection 620.


In one embodiment, a DC offset measurement 630 allows DC offset removal 640 (shown as an element labeled with a Greek letter sigma). The DC offset removal 620 is to allow the compander 650 to be as efficient as possible. In one embodiment, the DC offset is recorded with the FLAC data, as meta data 670, so that the server can re-add the DC offset after expanding (un-companding) the data.


The compander 650 is used to reduce the uploaded data size, removing non-essential values from the data stream.


The compressed data is then encoded, in one embodiment. In one embodiment, free lossless audio codec (FLAC) encoder 660 is used to encode the data. In another embodiment, another lossless compression algorithm may be used, such as MPEG-4 ALS. In other embodiments, alternative encoding may be used. In one embodiment a lossy compression, such as a variant of MP3 may be used. In such an embodiment, the compression may be tuned for the data content so that the loss is minimal.


The FLAC data is stored in a memory 680 and then uploaded by uploader 690, to server via a network 695. In one embodiment, the uploader 690 uploads bursts of data. In one embodiment, the upload interval is specified by the cloud servers. In another embodiment, the uploader 690 uploads data when a certain amount of data is accumulated. This may result in slower uploads for data with a lower sample rate.


In one embodiment, the digital portion of the system runs in firmware on a processor.



FIG. 7 is a flowchart of one embodiment of data collection through upload to the server. The process starts at block 710. At block 715, insulated piezoelectric sensors are used to capture charge corresponding to motion. This data recording is sufficiently sensitive so that the data reflects not only body movements, but also the movement of the chest in breathing, as well as the movement of the rib cage in heart beats. The sensor is sensitive enough that it can record, and the AI-enabled system can identify, snoring based on the vibration of the user's throat, which is detected by the sensor.


The system, at block 720, uses an insulated cable to send the charge data to the receiver, in one embodiment. Because the data is very precise even a small amount of noise can reduce the precision sufficiently to create an issue. Therefore, the data from the time the sensor detects it, until it is converted to a voltage, is run through an insulated system.


The insulated cable connects the data to insulated plugs, where the data is received at block 725, in one embodiment.


A charge-to-voltage converter is used to obtain a voltage corresponding to the charge data, reflecting the motion sensed by the sensors, at block 730. Optionally, at block 735, a conditioning amplifier may be used to adjust the voltage range for the A-to-D converter.


At block 740, the A-to-D converter converts the voltages to a digital signal. In one embodiment, the insulation may extend to the A-to-D converter. In another embodiment, once the signal is converted to a voltage, the signal is more robust, and the path need no longer be fully insulated.


At block 745, the data is compandered, in one embodiment.


At block 750, the data is encoded for transmission. In one embodiment, the encoding uses a lossless encoding algorithm. In one embodiment, a FLAC encoding is used. In one embodiment, this allows the use of an audio CODEC for the encoding.


At block 755, in one embodiment the data is stored in buffer. In one embodiment, at block 760 the data is sent in bursts to the server. In another embodiment, the data may be sent continuously. In another embodiment, the data may be sent periodically. The process then ends at block 770.


Of course, though this is shown as a flowchart, in one embodiment it is implemented as an interrupt-driven system, such that the device state is changed when a state detection system identifies a change of the state. Additionally, the ordering of state checking is arbitrary.


In the foregoing specification, the invention has been described with reference to specific exemplary embodiments thereof. It will, however, be evident that various modifications and changes may be made thereto without departing from the broader spirit and scope of the invention as set forth in the appended claims. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims
  • 1. A hardware sleep sensor system comprising: a sensor outputting sensor data corresponding to a motion;a converter to convert the sensor data to digital data;a downsampler configured to receive a down-sampling rate calculated based on a level of use of the sleep sensor system, wherein the level of use is one of: not in use level, limited analysis use level, and full use level;the downsampler further configured to down-sample an output of the converter using the down-sampling rate; andan uploader to upload the down-sampled digital data to a server for processing.
  • 2. The system of claim 1, wherein the sensor is a piezo sensor outputting charge data corresponding to the motion.
  • 3. The system of claim 1, further comprising: an insulated portion of the hardware sleep sensor system, the insulated portion including the sensor, a plug, and the converter.
  • 4. The system of claim 3, wherein the insulated portion is insulated using a metal enclosure.
  • 5. The system of claim 1, wherein the sensor is a piezo sensor, and the system further comprising: an electrostatic discharge protector coupled to a cable coupling the piezo sensor to a receiver.
  • 6. The system of claim 1, wherein the converter is an analog-to-digital converter that a lossless algorithm to convert the sensor data to the digital data.
  • 7. The system of claim 6, wherein the sensor data is charge data, and the system further comprising: a charge to voltage converter to convert the charge data from the sensor into voltage data; andthe analog-to-digital converter to convert the voltage data to the digital data.
  • 8. The system of claim 7, wherein the lossless algorithm is FLAC (free lossless compression), and the analog-to-digital converter is an audio codec.
  • 9. The system of claim 1, comprising: the sensor comprising two piezo sensors; anda two-channel analog-to-digital converter to provide concurrent sampling of data from the two piezo sensors.
  • 10. The system of claim 1, wherein: the not in use level is defined as no user data being measured;the limited analysis use level is defined as measuring: sleep state, heart rate and breathing rate; andthe full use level is defined as measuring the sleep state, the heart rate, and the breathing rate, and further measuring one or more of: snore detection, respiratory events, and heart rate variability.
  • 11. The system of claim 1, wherein the down-sampling rate is 0.5 Hz to 5 Hz for the not in use level; is 30 Hz to 100 Hz for the limited analysis use level; and is 100 Hz to 500 Hz for the full use level.
  • 12. The hardware sleep sensor system of claim 1, wherein: the sensor outputting sensor data comprises two piezo sensors outputting charge data corresponding to the motion on a sleep surface;an insulated portion of a receiver to receive the charge data from the two piezo sensors, the insulated portion coupled to the two piezo sensors via an insulated cable, and an insulated plug, the insulated cable including an electrostatic discharge protector;the converter to convert sensor data comprises a charge to voltage converter for converting the charge data to voltage data, the charge to voltage converter on the insulated portion of the receiver, and a two-channel audio codec digital-to-analog converter to concurrently sample the charge data from the two piezo sensors, and to convert the voltage data to digital data using a lossless algorithm;the downsampler further configured to down-sample an output of the converter using the down-sampling rate.
  • 13. A method of monitoring a user with a sleep sensor system, the method comprising: receiving sensor data from a sensor, the sensor data corresponding to a motion of the user;converting the sensor data to digital data;receiving a down-sampling rate calculated based on a level of use of the sleep sensor system, wherein the level of use is one of: not in use, limited analysis use, and full use;down-sampling the digital data using the down-sampling rate; anduploading the down-sampled digital data to a server for processing.
  • 14. The method of claim 13, wherein the sensor data is received from a piezo sensor outputting charge data corresponding to the motion.
  • 15. The method of claim 13, wherein a lossless algorithm is used to convert the sensor data to the digital data.
  • 16. The method of claim 15, wherein the sensor data is charge data, and the method further comprises: converting the charge data from the sensor into voltage data; andconverting the voltage data to the digital data.
  • 17. The method of claim 15, wherein the lossless algorithm is an audio codec.
  • 18. The method of claim 13, further comprising: receiving sensor data an additional sensor; andusing a two-channel analog-to-digital converter to provide concurrent sampling of the data from the sensor and the additional sensor.
  • 19. The method of claim 13, wherein: the not in use level is defined as no user data being measured;the limited analysis use level is defined as measuring: sleep state, heart rate and breathing rate; andthe full use level is defined as measuring the sleep state, the heart rate, and the breathing rate, and further measuring one or more of: snore detection, respiratory events, and heart rate variability.
  • 20. The method of claim 13, wherein the down-sampling rate is 0.5 Hz to 5 Hz for the not in use level; is 30 Hz to 100 Hz for the limited analysis use level; and is 100 Hz to 500 Hz for the full use level.
  • 21. A hardware sleep sensor system comprising: two piezo sensors for outputting charge data corresponding to motion on a sleep surface;an insulated portion of a receiver to receive the charge data from the two piezo sensors, the insulated portion coupled to the two piezo sensors via an insulated cable, and an insulated plug, the insulated cable including an electrostatic discharge protector;a charge to voltage converter for converting the charge data to voltage data, the charge to voltage converter on the insulated portion of the receiver;a two-channel audio codec digital-to-analog converter to concurrently sample the charge data from the two piezo sensors, and to convert the voltage data to digital data using a lossless algorithm;a downsampler to down-sample an output of the digital-to-analog converter; andan uploader to upload the down-sampled output to a server for processing.
RELATED APPLICATION

The present application is a continuation of U.S. patent application Ser. No. 16/601,561, filed Oct. 14, 2019, issuing as U.S. Pat. No. 11,471,097 on Oct. 18, 2022, which application claims priority to U.S. Provisional Patent Application 62/745,976 filed on Oct. 15, 2018. The present application also claims priority to U.S. Provisional Patent Application No. 62/745,978 (8689P232Z) and U.S. Provisional Application No. 62/745,984 (8689P233Z) both filed on Oct. 15, 2019 and incorporates all three of those applications by reference in their entirety.

US Referenced Citations (281)
Number Name Date Kind
2082843 Mathez Jun 1937 A
3541781 Bloom Nov 1970 A
3683933 Mansfield Aug 1972 A
3798889 Chadwick Mar 1974 A
4228806 Lidow Oct 1980 A
4297685 Brainard, II Oct 1981 A
4322609 Kato Mar 1982 A
4573804 Kavoussi et al. Mar 1986 A
4788533 Mequignon Nov 1988 A
4848360 Palsgard et al. Jul 1989 A
4858609 Cole Aug 1989 A
4982738 Griebel Jan 1991 A
5008865 Shaffer et al. Apr 1991 A
5047930 Martens et al. Sep 1991 A
5168759 Bowman Dec 1992 A
5275159 Griebel Jan 1994 A
5335657 Terry et al. Aug 1994 A
5458105 Taylor et al. Oct 1995 A
5545192 Czeisler et al. Aug 1996 A
5562106 Heeke et al. Oct 1996 A
5671733 Raviv et al. Sep 1997 A
5844996 Enzmann et al. Dec 1998 A
5868647 Belsole Feb 1999 A
5928133 Halyak Jul 1999 A
5961447 Raviv et al. Oct 1999 A
6014682 Stephen et al. Jan 2000 A
6045514 Raviv et al. Apr 2000 A
6231527 Sol May 2001 B1
6239706 Yoshiike et al. May 2001 B1
6350275 Vreman et al. Feb 2002 B1
6361508 Johnson et al. Mar 2002 B1
6468234 Van et al. Oct 2002 B1
6547728 Cornuejols Apr 2003 B1
6556222 Narayanaswami Apr 2003 B1
6834436 Townsend et al. Dec 2004 B2
6888779 Mollicone et al. May 2005 B2
6928031 Kanevsky et al. Aug 2005 B1
6963271 Fyffe Nov 2005 B1
7006650 Wild Feb 2006 B1
7041049 Raniere May 2006 B1
7106662 Acker, Jr. Sep 2006 B1
7139342 Phanse Nov 2006 B1
7153278 Ono et al. Dec 2006 B2
7280439 Shaddox Oct 2007 B1
7366572 Heruth et al. Apr 2008 B2
7513003 Mossbeck Apr 2009 B2
7559903 Moussavi et al. Jul 2009 B2
7572225 Stahmann et al. Aug 2009 B2
7652581 Gentry et al. Jan 2010 B2
7841987 Sotos et al. Nov 2010 B2
7862226 Bracher et al. Jan 2011 B2
7868757 Radivojevic et al. Jan 2011 B2
7914468 Shalon et al. Mar 2011 B2
7974849 Begole et al. Jul 2011 B1
8179270 Rai et al. May 2012 B2
8193941 Wolfe et al. Jun 2012 B2
8398546 Pacione et al. Mar 2013 B2
8407835 Connor Apr 2013 B1
8475339 Hwang et al. Jul 2013 B2
8482418 Harman Jul 2013 B1
8577448 Bauer et al. Nov 2013 B2
8680974 Meiertoberens et al. Mar 2014 B2
8738925 Park et al. May 2014 B1
8892036 Causey et al. Nov 2014 B1
8909357 Rawls-Meehan Dec 2014 B2
8942719 Hyde et al. Jan 2015 B1
9060735 Yang et al. Jun 2015 B2
9161719 Tsutsumi et al. Oct 2015 B2
9257029 Hendrick et al. Feb 2016 B1
9448536 Kahn et al. Sep 2016 B1
9474876 Kahn et al. Oct 2016 B1
9594354 Kahn et al. Mar 2017 B1
9675268 Bauer et al. Jun 2017 B2
9844336 Zigel et al. Dec 2017 B2
10004452 Kazem-Moussavi et al. Jun 2018 B2
10207075 Kahn et al. Feb 2019 B1
10252058 Fuerst Apr 2019 B1
10335060 Kahn et al. Jul 2019 B1
10842968 Kahn et al. Nov 2020 B1
11100922 Mutagi et al. Aug 2021 B1
20010049482 Pozos et al. Dec 2001 A1
20020080035 Youdenko Jun 2002 A1
20020100477 Sullivan et al. Aug 2002 A1
20020124848 Sullivan et al. Sep 2002 A1
20030095476 Mollicone et al. May 2003 A1
20030204412 Brier Oct 2003 A1
20030227439 Lee et al. Dec 2003 A1
20030231495 Searfoss Dec 2003 A1
20040034289 Teller et al. Feb 2004 A1
20040049132 Barron et al. Mar 2004 A1
20040071382 Rich et al. Apr 2004 A1
20040111039 Minamiura et al. Jun 2004 A1
20040133081 Teller et al. Jul 2004 A1
20040210155 Takemura et al. Oct 2004 A1
20040218472 Narayanaswami et al. Nov 2004 A1
20050012622 Sutton Jan 2005 A1
20050043645 Ono et al. Feb 2005 A1
20050075116 Laird et al. Apr 2005 A1
20050076715 Kuklis et al. Apr 2005 A1
20050143617 Auphan Jun 2005 A1
20050154330 Loree et al. Jul 2005 A1
20050190065 Ronnholm Sep 2005 A1
20050236003 Meader Oct 2005 A1
20050237479 Rose Oct 2005 A1
20050245793 Hilton et al. Nov 2005 A1
20050283039 Cornel Dec 2005 A1
20050288904 Warrior et al. Dec 2005 A1
20060017560 Albert Jan 2006 A1
20060025299 Miller et al. Feb 2006 A1
20060064037 Shalon et al. Mar 2006 A1
20060097884 Jang et al. May 2006 A1
20060136018 Lack et al. Jun 2006 A1
20060150734 Mimnagh-Kelleher et al. Jul 2006 A1
20060252999 Devaul et al. Nov 2006 A1
20060266356 Sotos et al. Nov 2006 A1
20060279428 Sato et al. Dec 2006 A1
20060293602 Clark Dec 2006 A1
20060293608 Rothman et al. Dec 2006 A1
20070016091 Butt et al. Jan 2007 A1
20070016095 Low et al. Jan 2007 A1
20070093722 Noda et al. Apr 2007 A1
20070100666 Stivoric et al. May 2007 A1
20070129644 Richards et al. Jun 2007 A1
20070139362 Colton et al. Jun 2007 A1
20070191692 Hsu et al. Aug 2007 A1
20070239225 Saringer Oct 2007 A1
20070250286 Duncan et al. Oct 2007 A1
20070251997 Brown et al. Nov 2007 A1
20070287930 Sutton Dec 2007 A1
20080062818 Plancon et al. Mar 2008 A1
20080109965 Mossbeck May 2008 A1
20080125820 Stahmann et al. May 2008 A1
20080169931 Gentry et al. Jul 2008 A1
20080191885 Loree et al. Aug 2008 A1
20080234785 Nakayama et al. Sep 2008 A1
20080243014 Moussavi et al. Oct 2008 A1
20080269625 Halperin Oct 2008 A1
20080275348 Catt et al. Nov 2008 A1
20080275349 Halperin et al. Nov 2008 A1
20080289637 Wyss Nov 2008 A1
20080319277 Bradley Dec 2008 A1
20090030767 Morris et al. Jan 2009 A1
20090048540 Otto et al. Feb 2009 A1
20090069644 Hsu et al. Mar 2009 A1
20090071810 Hanson et al. Mar 2009 A1
20090082699 Bang et al. Mar 2009 A1
20090094750 Oguma et al. Apr 2009 A1
20090105785 Wei et al. Apr 2009 A1
20090121826 Song et al. May 2009 A1
20090128487 Langereis et al. May 2009 A1
20090143636 Mullen et al. Jun 2009 A1
20090150217 Luff Jun 2009 A1
20090177327 Turner et al. Jul 2009 A1
20090203970 Fukushima et al. Aug 2009 A1
20090203972 Heneghan et al. Aug 2009 A1
20090207028 Kubey et al. Aug 2009 A1
20090209839 Ochs et al. Aug 2009 A1
20090227888 Salmi et al. Sep 2009 A1
20090264789 Molnar et al. Oct 2009 A1
20090320123 Yu et al. Dec 2009 A1
20100010330 Rankers et al. Jan 2010 A1
20100010565 Lichtenstein Jan 2010 A1
20100036211 La et al. Feb 2010 A1
20100061596 Mostafavi et al. Mar 2010 A1
20100075807 Hwang et al. Mar 2010 A1
20100079291 Kroll et al. Apr 2010 A1
20100079294 Rai et al. Apr 2010 A1
20100083968 Wondka et al. Apr 2010 A1
20100094139 Brauers et al. Apr 2010 A1
20100094148 Bauer et al. Apr 2010 A1
20100100004 Van Someren Apr 2010 A1
20100102971 Virtanen et al. Apr 2010 A1
20100152543 Heneghan et al. Jun 2010 A1
20100152546 Behan et al. Jun 2010 A1
20100217146 Osvath Aug 2010 A1
20100256512 Sullivan Oct 2010 A1
20100283618 Wolfe et al. Nov 2010 A1
20100331145 Lakovic et al. Dec 2010 A1
20110015467 Dothie et al. Jan 2011 A1
20110015495 Dothie et al. Jan 2011 A1
20110018720 Rai et al. Jan 2011 A1
20110046498 Klap et al. Feb 2011 A1
20110054279 Reisfeld et al. Mar 2011 A1
20110058456 Van et al. Mar 2011 A1
20110090226 Sotos et al. Apr 2011 A1
20110105915 Bauer et al. May 2011 A1
20110137836 Kuriyama et al. Jun 2011 A1
20110160619 Gabara Jun 2011 A1
20110190594 Heit et al. Aug 2011 A1
20110199218 Caldwell et al. Aug 2011 A1
20110230790 Kozlov Sep 2011 A1
20110245633 Goldberg et al. Oct 2011 A1
20110295083 Doelling et al. Dec 2011 A1
20110302720 Yakam et al. Dec 2011 A1
20110304240 Meitav et al. Dec 2011 A1
20120004749 Abeyratne et al. Jan 2012 A1
20120083715 Yuen et al. Apr 2012 A1
20120232414 Mollicone et al. Sep 2012 A1
20120243379 Balli Sep 2012 A1
20120253220 Rai et al. Oct 2012 A1
20120296156 Auphan Nov 2012 A1
20130012836 Crespo et al. Jan 2013 A1
20130018284 Kahn et al. Jan 2013 A1
20130023214 Wang et al. Jan 2013 A1
20130053653 Cuddihy et al. Feb 2013 A1
20130053656 Mollicone et al. Feb 2013 A1
20130060306 Colbauch Mar 2013 A1
20130144190 Bruce et al. Jun 2013 A1
20130184601 Zigel et al. Jul 2013 A1
20130197857 Lu et al. Aug 2013 A1
20130204314 Miller et al. Aug 2013 A1
20130208576 Loree et al. Aug 2013 A1
20130283530 Main et al. Oct 2013 A1
20130286793 Umamoto Oct 2013 A1
20130289419 Berezhnyy et al. Oct 2013 A1
20130300204 Partovi Nov 2013 A1
20130310658 Ricks et al. Nov 2013 A1
20130344465 Dickinson et al. Dec 2013 A1
20140005502 Klap et al. Jan 2014 A1
20140051938 Goldstein et al. Feb 2014 A1
20140085077 Luna et al. Mar 2014 A1
20140135955 Burroughs May 2014 A1
20140171815 Yang et al. Jun 2014 A1
20140200691 Lee et al. Jul 2014 A1
20140207292 Ramagem et al. Jul 2014 A1
20140218187 Chun et al. Aug 2014 A1
20140219064 Filipi et al. Aug 2014 A1
20140232558 Park et al. Aug 2014 A1
20140256227 Aoki et al. Sep 2014 A1
20140259417 Nunn et al. Sep 2014 A1
20140259434 Nunn et al. Sep 2014 A1
20140273858 Panther et al. Sep 2014 A1
20140276227 Perez Sep 2014 A1
20140288878 Donaldson Sep 2014 A1
20140306833 Ricci Oct 2014 A1
20140350351 Halperin et al. Nov 2014 A1
20140371635 Shinar et al. Dec 2014 A1
20150015399 Gleckler et al. Jan 2015 A1
20150068069 Tran et al. Mar 2015 A1
20150073283 Van et al. Mar 2015 A1
20150085622 Carreel et al. Mar 2015 A1
20150094544 Spolin et al. Apr 2015 A1
20150098309 Adams et al. Apr 2015 A1
20150101870 Gough et al. Apr 2015 A1
20150136146 Hood et al. May 2015 A1
20150141852 Dusanter et al. May 2015 A1
20150148621 Sier May 2015 A1
20150148871 Maxik et al. May 2015 A1
20150164238 Benson et al. Jun 2015 A1
20150164409 Benson et al. Jun 2015 A1
20150164438 Halperin et al. Jun 2015 A1
20150164682 Remmers et al. Jun 2015 A1
20150173671 Paalasmaa et al. Jun 2015 A1
20150178362 Wheeler Jun 2015 A1
20150190086 Chan et al. Jul 2015 A1
20150220883 Lingg et al. Aug 2015 A1
20150233598 Shikii et al. Aug 2015 A1
20150238139 Raskin et al. Aug 2015 A1
20150265903 Kolen et al. Sep 2015 A1
20150289802 Thomas et al. Oct 2015 A1
20150320588 Connor Nov 2015 A1
20150333950 Johansson Nov 2015 A1
20150346824 Chen Dec 2015 A1
20150351694 Shimizu et al. Dec 2015 A1
20160015315 Auphan et al. Jan 2016 A1
20160045035 Van Erlach Feb 2016 A1
20160217672 Yoon et al. Jul 2016 A1
20160262693 Sheon Sep 2016 A1
20160287869 Errico Oct 2016 A1
20170003666 Nunn et al. Jan 2017 A1
20170020756 Hillenbrand et al. Jan 2017 A1
20170188938 Toh et al. Jul 2017 A1
20180049701 Raisanen Feb 2018 A1
20180103770 Nava et al. Apr 2018 A1
20180338725 Shan et al. Nov 2018 A1
20190021675 Gehrke et al. Jan 2019 A1
20190044380 Lausch et al. Feb 2019 A1
20190132570 Chen et al. May 2019 A1
20190156296 Lu et al. May 2019 A1
20190190992 Warrick Jun 2019 A1
20190201270 Sayadi et al. Jul 2019 A1
Foreign Referenced Citations (15)
Number Date Country
2003203967 Nov 2004 AU
377738 Jan 1964 CH
668349 Dec 1988 CH
697528 Nov 2008 CH
4101471 Jul 1992 DE
19642316 Apr 1998 DE
1139187 Oct 2001 EP
08-160172 Jun 1996 JP
2007-132581 May 2007 JP
10-2009-0085403 Aug 2009 KR
10-2010-0022217 Mar 2010 KR
9302731 Feb 1993 WO
2008038288 Apr 2008 WO
2009099292 Aug 2009 WO
2011141840 Nov 2011 WO
Non-Patent Literature Citations (38)
Entry
“NPL—EasySense LTD”, archive.org, accessed: Jan. 7, 2019, published: Nov. 27, 2006.
Acligraphy, From Wikipedia, the free encyclopedia, downloaded at: http://en.wikipedia.org/wiki/Actigraphy on Apr. 24, 2014, 4 pages.
Advisory Action, U.S. Appl. No. 16/601,561, Feb. 1, 2022, 3 pages.
Campbell, Appleinsider, “Apple buys sleep tracking firm Beddit” May 9, 2017. Retrieved from https://appleinsider.com/articles/17/05/09/apple-buys-sleep-tracking-firm-beddit (Year: 2017).
Crist, CNET “Samsung introduces SleepSense” Sep. 3, 2015. Retrieved from https://www.cnet.com/reviews/samsung-sleepsense-preview (Year: 2015).
Daniel et al., “Activity Characterization from Actimetry Sensor Data for Sleep Disorders Diagnosis”, Sep. 2008, 10 pages.
Desai, Rajiv, “The Sleep”, Mar. 17, 2011, Educational Blog, 82 pages.
Final Office Action, U.S. Appl. No. 16/601,561, Nov. 5, 2021, 21 pages.
Fitbit Product Manual, “Fitbit Product Manual”, available online at <http://www.filtbit.com/manual>, Mar. 29, 2010, pp. 1-20.
Haughton Mifflin, “Estimate”, The American Heritage dictionary of the English language (5th ed.), Jul. 24, 2017, 2 pages.
How BodyMedia FIT Works, <http://www.bodymedia.com/Shop/Learn-More/How-it-works>, accessed Jun. 17, 2011, 2 pages.
Internet Archive, Withings “Sleep Tracking Mat” Nov. 22, 2018. Retrieved from https://web.archive.org/web/20181122024547/https://www.withings.com/us/en/sleep (Year: 2018).
Jaines, Kira, “Music to Help You Fall Asleep,” <http://www.livestrong.com/article/119802-music-fall-sleep/>, May 10, 2010, 2 pages.
JETLOG Reviewers Guide, <http://www.jetlog.com/fileadmin/Presse_us/24x7ReviewersGuide.pdf>, 2009, 5 pages.
Leeds, Joshua, “Sound-Remedies.com: Sonic Solutions for Health, Learning & Productivity,” <http://www.sound-remedies.com/ammusforslee.html>, Accessed May 23, 2013, 2 pages.
Lichstein, et al., “Actigraphy Validation with Insomnia”, SLEEP, vol. 29, No. 2, 2006, pp. 232-239.
Liden et al., “Characterization and Implications of the Sensors Incorporated into the SenseWear(TM) Armband for Energy Expenditure and Activity Detection”, 2011, 7 pages.
Mattila et al., “A Concept for Personal Wellness Management Based on Activity Monitoring,” Pervasive Computing Technologies for Healthcare, 2008.
Notice of Allowance, U.S. Appl. No. 16/601,561, Jun. 17, 2022, 13 pages.
Patel, et al., Validation of Basis Science Advanced Sleep Analysis, Estimation of Sleep Stages and Sleep Duration, Basis Science, San Francisco, CA, Jan. 2014, 6 pages.
Pires, P. D. C. Activity Characterization from Actimetry Sensor Data for Sleep Disorders Diagnosis, Universidade T ecnica de Lisboa, Sep. 2008, 10 pages.
Pollak et al., “How Accurately Does Wrist Actigraphy Identify the States of Sleep and Wakefulness?”, Sleep, vol. 24, No. 8, 2001, pp. 957-965.
Power Nap, <en.wikipedia.org/wiki/Power.sub.-nap>, Last Modified Sep. 20, 2012, 4 pages.
PowerNap, “iPhone App”, available online at <http://forums.precentral.net/webos-apps-software/223091-my-second-app-powernap-out-app-catalog-nap-timer.html>, Jan. 6, 2010, 10 pages.
Rechtschaffen et al., Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects, 1968, 57 pages.
Sara Mednick, <en.wikipedia.org/wiki/Sara.sub.-Mednick>, Last Modified Sep. 12, 2012, 2 pages.
Schulz et al. “Phase shift in the REM sleep rhythm.” Pflugers Arch. 358, 1975, 10 pages.
Schulz et al. “The REM-NREM sleep cycle: Renewal Process or Periodically Driven Process?.” Sleep, 1980, pp. 319-328.
Sleep Debt, <en.wikipedia.org/wiki/Sleep.sub.-debt>, Last Modified Aug. 25, 2012, 3 pages.
Sleep Inertia, <en.wikipedia.org/wiki/Sleep_inertia>, Last Modified Sep. 12, 2012, 2 pages.
Sleep, <en.wikipedia.org/wiki/Sleep.sub.-stages#Physiology>, Last Modified Oct. 5, 2012, 21 pages.
Slow Wave Sleep, <en.wikipedia.org/wiki/Slow-wave.sub.-sleep>, Last Modified Jul. 22, 2012, 4 pages.
Sunseri et al., “The SenseWear (TM) Armband as a Sleep Detection Device,” available online at <http://sensewear.bodymedia.com/SenseWear-Sludies/SW-Whilepapers/The-SenseWear-armband-as-a-Sleep-Delection-Device>, 2005, 9 pages.
Wikipedia, “David.sub Dinges”, available online at <en.wikipedia.org/wiki/David.sub_Dinges>, Sep. 12, 2012, 2 pages.
Yassourdidis et al. “Modelling and Exploring Human Sleep with Event History Analysis.” Journal of Sleep Research, 1999, pp. 25-36.
Ding, F., et al., “Polysomnographic validation of an under-mattress monitoring device in estimating sleep architecture and obstructive sleep apnea in adults,” Sleep Medicine, vol. 96, Apr. 2022, pp. 20-27.
Choudhary, S. and Choudhary, S, Sleep Effects on Breathing and Respiratory Diseases, Review Article, Oct.-Dec. 2009, 117-122 pages, vol. 26, Issue 4, Department of Pulmonary Medicine, Sleep Medicine, Critical Care, Shree Ramjevan Choudhary Memorial Hospital and Research Centre, Nagpur—02, India, 6 pages.
Verbraecken, J, Applications of Evolving Technologies in Sleep Medicine, Article, Dec. 2013, 443-455 pages, vol. 9, No. 6, Dept of Pulmonary Medicine and Multidisciplinary Sleep Disorders Centre, Antwerp University Hospital, Antwerp, Belgium, 14 pages.
Provisional Applications (3)
Number Date Country
62745976 Oct 2018 US
62745984 Oct 2018 US
62745978 Oct 2018 US
Continuations (1)
Number Date Country
Parent 16601561 Oct 2019 US
Child 18047258 US