Activity identification using an optical heart rate monitor

Information

  • Patent Grant
  • 10285626
  • Patent Number
    10,285,626
  • Date Filed
    Friday, December 12, 2014
    9 years ago
  • Date Issued
    Tuesday, May 14, 2019
    5 years ago
Abstract
An electronic device that can be worn by a user can include a processing device and one or more optical heart rate monitors operatively connected to the processing device. The processing device can be adapted to receive a OHRM signal from at least one optical heart rate monitor. The OHRM signal includes one or more motion artifacts that are produced by a physical activity of the user. The processing device can be adapted to analyze the OHRM signal to determine the physical activity of the user.
Description
TECHNICAL FIELD

The present invention relates generally to electronic devices, and more particularly to wearable electronic devices. Still more particularly, the present invention relates to determining a physical activity based on a signal received from at least one optical heart rate monitor.


BACKGROUND

Portable electronic devices can be used for performing a wide variety of tasks, and in some situations the electronic device can be worn on the body of a user. For example, a portable electronic device can be worn by a user on his or her wrist, arm, ankle, or leg. One example of such an electronic device is a wrist-worn activity monitor. The activity monitor can include a heart rate monitor, a position sensor (e.g., gyroscope), and/or a motion sensor (e.g., accelerometer). The activity monitor can determine the type of physical activity based on the signals received from the heart rate monitor and the sensor(s).


Some activities, however, involve little or no limb motion during the performance of the physical activity. For example, a user's arms can remain substantially still when the user is bicycling, walking or running while pushing a stroller, exercising on an elliptical trainer or stair machine while holding the handles or side railings, and performing low-impact activities such as push-ups, squats, or sit-ups. In these situations, it can be difficult, if not impossible, for a wrist-worn activity monitor to determine the type of physical activity the user is performing. The wrist-worn activity monitor may be unable to provide information to the user about the user's physical condition or his or her performance during the physical activity. For example, the wrist-worn activity monitor may not be able to present the user with the number of steps taken by the user or the number of calories expended during the physical activity.


SUMMARY

A signal produced by an optical heart rate monitor (OHRM) can include motion artifacts or noise that are introduced into the signal during physical activity. For example, motion of the body part wearing the OHRM, motion between the OHRM and the skin, and variations in blood flow caused by body movement (e.g., a physical activity of the user) can produce motion artifacts or noise in the signal produced by the OHRM. Embodiments described herein determine the type of physical activity performed by a user by analyzing the OHRM signal that includes one or more motion artifacts.


In one aspect, an electronic device can include a processing device and one or more OHRMs operatively connected to the processing device. The processing device may be adapted to receive an OHRM signal from at least one OHRM when the user performs a physical activity. The OHRM signal includes one or more motion artifacts that are produced by the physical activity, and the processing device can be adapted to analyze the OHRM signal to determine the physical activity of the user.


In another aspect, a method for determining a physical activity of a user wearing an electronic device that includes an OHRM can include receiving an OHRM signal from the OHRM and analyzing the OHRM signal to determine the physical activity of the user. The OHRM signal includes one or more motion artifacts that are produced while the user performs the physical activity.


In another aspect, a system can include an OHRM, one or more motion and/or position sensors, and a processing device operatively connected to the OHRM and the motion and/or position sensor(s). The processing device is adapted to receive an OHRM signal from the OHRM. The OHRM signal includes one or more motion artifacts that are produced by the physical activity of the user. The processing device can also be adapted to receive a sensor signal from at least one motion and/or position sensor. The processing device analyzes the OHRM signal and the sensor signal to determine a physical activity performed by the user. Additionally or alternatively, information regarding the activity can be provided to the user. For example, data such as a heart rate, the number of steps taken, cadence information, the intensity of the activity, calorie consumption, and/or the user's speed can be provided to the user.


In yet another aspect, an electronic device includes an OHRM. The electronic device can be calibrated to determine a physical activity of a user by receiving an OHRM signal that includes one or more motion artifacts when a user performs a particular physical activity, receiving an activity identifier, and associating the activity identifier to the OHRM signal. Subsequent OHRM signals can then be correlated to an activity based on the associated activity identifier, and the identified activity may be displayed or provided to the user.





BRIEF DESCRIPTION OF THE DRAWINGS

Embodiments of the invention are better understood with reference to the following drawings. The elements of the drawings are not necessarily to scale relative to each other. Identical reference numerals have been used, where possible, to designate identical features that are common to the figures



FIG. 1 is a perspective view of one example of a wearable electronic device that includes one or more optical heart rate monitors;



FIG. 2 is an illustrative block diagram of the wearable electronic device 100 shown in FIG. 1;



FIG. 3 is a flowchart of a method for detecting a physical activity of a user wearing an electronic device that includes one or more optical heart rate monitors;



FIG. 4 depicts a first example of a photoplethysmograph signal with motion artifacts and a filtered photoplethysmograph signal;



FIG. 5 illustrates a second example of a photoplethysmograph signal with motion artifacts and a filtered photoplethysmograph signal; and



FIG. 6 is a flowchart of a method for calibrating a wearable electronic device to determine one or more activities of a user.





DETAILED DESCRIPTION

Embodiments described herein provide a wearable electronic device that includes one or more optical heart rate monitors (OHRM). A signal received from at least one OHRM can include one or more motion artifacts or noise that is generated by movement of the user. Motion by the body part wearing the OHRM, motion between the OHRM and the skin, and variations in blood flow caused by body movement are example functions that can produce motion artifacts or noise in the signal output by an OHRM.


Embodiments described herein determine a physical activity of a user by analyzing an OHRM signal received from one or more OHRMs. The OHRM signal includes one or more motion artifacts that is produced by a physical activity of the user. One or more characteristics of the OHRM signal may be analyzed to determine the physical activity. For example, peak amplitudes, changes in amplitude, the distances between the peak amplitudes, time variations between peak amplitudes, the shape of the OHRM signal, and/or the frequency or frequency variations of the signal are characteristics of the OHRM signal that can be analyzed to identify the physical activity of the user.


In some embodiments, a signal produced by other types of sensors can be included in the analysis to determine the physical activity. As one example, a sensor signal from one or more motion and/or position sensors can be received and analyzed when determining a physical activity of the user. For example, when a user is mowing the lawn, a signal from an OHRM will include motion artifacts produced by the walking and/or pushing of the lawn mower. The OHRM signal can be analyzed to determine the user is mowing the lawn. Additionally, a signal from a gyroscope can detect turning position changes that indicate the user is mowing. Velocity determined from a signal received from a global positioning sensor can be consistent with the user's lawn mowing activity.


Referring now to FIG. 1, there is shown a perspective view of one example of a wearable electronic device that can include one or more optical heart rate monitors. Embodiments described herein include an electronic device 100 that is worn on a wrist of a user. But other embodiments can implement the electronic device differently, such as, for example, as a smart telephone, a gaming device, a digital music player, headphones or ear buds, a device that provides time, a health assistant, a fitness monitor, a medical device, and any other wearable electronic device. Additionally, the electronic device can be worn on any limb or other suitable body part (e.g., the head).


The wearable electronic device 100 includes an enclosure 102 at least partially surrounding a display 104 and one or more buttons 106 or input devices. The enclosure 102 can form an outer surface or partial outer surface and protective case for the internal components of the electronic device 100, and may at least partially surround the display 104. The enclosure 102 can be formed of one or more components operably connected together, such as a front piece and a back piece. Alternatively, the enclosure 102 can be formed of a single piece operably connected to the display 104.


The display 104 can be implemented with any suitable technology, including, but not limited to, a multi-touch sensing touchscreen that uses liquid crystal display (LCD) technology, light emitting diode (LED) technology, organic light-emitting display (OLED) technology, organic electroluminescence (OEL) technology, or another type of display technology. At least one button 106 can take the form of a home button, which may be a mechanical button, a soft button (e.g., a button that does not physically move but still accepts inputs), an icon or image on a display or on an input region, and so on. Further, in some embodiments, the button or buttons 106 can be integrated as part of a cover glass of the electronic device.


The wearable electronic device 100 can be permanently or removably attached to a band 108. The band 108 can be made of any suitable material, including, but not limited to, leather, rubber or silicon, fabric, and ceramic. In the illustrated embodiment, the band is a wristband that wraps around the user's wrist. The wristband can include an attachment mechanism (not shown) to secure the band to the user's wrist. Example attachment mechanisms include, but are not limited to, a bracelet clasp, Velcro, and magnetic connectors. In other embodiments, the band can be elastic or stretchy such that it fits over the hand of the user and does not include an attachment mechanism.



FIG. 2 is an illustrative block diagram of the wearable electronic device 100 shown in FIG. 1. The electronic device 100 can include the display 104, one or more processing devices 200, memory 202, one or more input/output (I/O) devices 204, one or more sensors 206, a power source 208, a network communications interface 210, and one or more optical heart rate monitors (OHRM) 212. The display 104 may provide an image or video output for the electronic device 100. The display may also provide an input surface for one or more input devices, such as, for example, a touch sensing device and/or a fingerprint sensor. The display 104 may be substantially any size and may be positioned substantially anywhere on the electronic device 100.


The processing device 200 can control some or all of the operations of the electronic device 100. The processing device 200 can communicate, either directly or indirectly, with substantially all of the components of the electronic device 100. For example, a system bus or signal line 214 or other communication mechanisms can provide communication between the processing device(s) 200, the memory 202, the I/O device(s) 204, the sensor(s) 206, the power source 208, the network communications interface 210, and/or the OHRM(s) 212. The one or more processing devices 200 can be implemented as any electronic device capable of processing, receiving, or transmitting data or instructions. For example, the processing device(s) 200 can each be a microprocessor, a central processing unit (CPU), an application-specific integrated circuit (ASIC), a digital signal processor (DSP), or combinations of such devices. As described herein, the term “processing device” is meant to encompass a single processor or processing unit, multiple processors, multiple processing units, or other suitably configured computing element or elements.


The memory 202 can store electronic data that can be used by the electronic device 100. For example, a memory can store electrical data or content such as, for example, audio and video files, documents and applications, device settings and user preferences, timing signals, signals received from the one or more OHRMs and sensors, calibration signals, data structures or databases, and so on. The memory 202 can be configured as any type of memory. By way of example only, the memory can be implemented as random access memory, read-only memory, Flash memory, removable memory, or other types of storage elements, or combinations of such devices.


The one or more I/O devices 204 can transmit and/or receive data to and from a user or another electronic device. One example of an I/O device is button 106 in FIG. 1. The I/O device(s) 204 can include a display, a touch sensing input surface such as a trackpad, one or more buttons, one or more microphones or speakers, one or more ports such as a microphone port, and/or a keyboard.


The electronic device 100 may also include one or more sensors 206 positioned substantially anywhere on the electronic device 100. The sensor or sensors 206 may be configured to sense substantially any type of characteristic, such as but not limited to, images, pressure or force, position, motion, speed, light, touch, heat, biometric data, and so on. For example, the sensor(s) 206 may be an image sensor, a gyroscope, an accelerometer, a global positioning sensor, a heat sensor, a light or optical sensor, a pressure transducer, a magnetometer, a health monitoring sensor, and so on.


The power source 208 can be implemented with any device capable of providing energy to the electronic device 100. For example, the power source 208 can be one or more batteries or rechargeable batteries, or a connection cable that connects the remote control device to another power source such as a wall outlet. Additionally or alternatively, the power source 208 can include a wireless energy transfer device, such as an inductive energy receiver device.


The network communication interface 210 can facilitate transmission of data to or from other electronic devices. For example, a network communication interface can transmit electronic signals via a wireless and/or wired network connection. Examples of wireless and wired network connections include, but are not limited to, cellular, Wi-Fi, Bluetooth, IR, and Ethernet.


The one or more OHRMs 212 can each measure one or more physiological functions of the user wearing the wearable electronic device 100. Each OHRM can be implemented as any suitable optical heart rate monitor. For example, in one embodiment, at least one OHRM is a reflective or transmissive photoplethysmograph (PPG) sensor. Illustrative measurements that a PPG sensor can measure include heart rate, the relative blood flow through a body part of a user, heart rate variability, and blood volume pulse. As will be described in more detail later, an OHRM signal or signals that includes one or more motion artifacts produced by a physical activity of a user is received from at least one OHRM and analyzed to identify and/or classify the physical activity of the user.


In some embodiments, the electronic device 100 can communicate with an external electronic device 216 using connection 218. Connection 218 can be a wired or wireless connection. As one example, the connection can be a cellular, Wi-Fi, or Bluetooth connection. Alternatively, a physical connector cable can connect the wearable electronic device to the external electronic device. The external electronic device 216 can be any type of electronic device, such as a computing device. Example external electronic devices include, but are not limited to, a computer such as a laptop, a tablet computing device, a smart telephone, or another wearable electronic device.


The external electronic device can include a network communication interface 220 operably connected to a processing device 222 and a memory 224. The processing device 222 can control some or all of the operations of the external electronic device 216 through bus 226. Additionally or alternatively, the processing device 222 can control some or all of the operations of the wearable electronic device 100.


It should be noted that FIGS. 1 and 2 are illustrative only. In other examples, an electronic device may include fewer or more components than those shown in FIGS. 1 and 2. Additionally or alternatively, the wearable electronic device can be in communication with other external devices. For example, a wearable electronic device may be operatively connected to, or in communication with a separate display. As another example, a wearable electronic device can access one or more signals or data that is stored in a memory separate from the wearable electronic device.


Additionally or alternatively, in some embodiments one or more components shown in the electronic device 100 can instead be included in the external electronic device 216. For example, one or more sensors can be included in an external electronic device and the signal produced by the one or more sensors can be analyzed to determine a physical activity of the user. As one example, a user can wear the electronic device and carry a smart telephone at the same time. The wearable electronic device can be wirelessly paired to the smart telephone. A signal obtained from a global positioning sensor, a gyroscope, and/or an accelerometer in the smart telephone can be analyzed with an OHRM signal received from an OHRM in the electronic device to determine a physical activity of the user.


Referring now to FIG. 3, there is shown a flowchart of a method for detecting a physical activity of a user wearing an electronic device that includes one or more OHRMs. Initially, a signal that includes motion artifacts is received from at least one OHRM at block 300. For example, movement of the body part wearing the OHRM, motion between the OHRM and the skin, and variations in blood flow caused by body movement can produce motion artifacts in the signal produced by the OHRM.


Optionally, a signal can also be received from other types of sensors at block 300. In one embodiment, a sensor signal can be received from a motion sensor and/or a position sensor. Examples of motion and position sensors include, but are not limited to, a gyroscope, an accelerometer, a global positioning sensor, a rotation vector sensor, a proximity sensor, and/or a magnetometer.


Next, as shown in block 302, the OHRM signal is analyzed to determine a physical activity being performed by the user. For example, the OHRM signal can be analyzed by the processing device 200 and/or the processing device 222 shown in FIG. 2. The analysis can identify a physical activity of the user. One or more characteristics of the OHRM signal having one or more motion artifacts may be analyzed to identify the physical activity of the user. For example, peak amplitudes, changes in amplitude, the distances between the peak amplitudes, time variations between peak amplitudes, the shape of the OHRM signal, and/or the frequency of the signal are characteristics of the OHRM signal that can be analyzed at block 302.


Optionally, a signal received from one or more other sensors can be analyzed with the OHRM signal at block 302 to determine the physical activity of the user. For example, when a user is bicycling, a signal from an OHRM will include motion artifacts produced by body position changes occurring with each leg thrust. The OHRM signal can be analyzed to determine the user is bicycling. Additionally, a signal from a gyroscope can detect turning and changes in hand position that indicate the user is bicycling. The velocity determined from a signal received from a global positioning sensor can be consistent with the activity of bicycling. And if impacts or high frequency vibrations are detected by an accelerometer, it may be possible to classify the bicycling as mountain biking instead of bicycling on a road. Thus, one or more signals received from other types of sensors, such as motion and position sensors, can be used to determine the type of physical activity and/or to further classify the type of activity.


Next, as shown in block 304, one or more of the signals can be processed to provide the user with additional information regarding the physical activity and/or his or her performance. The one or more signals may include the OHRM signal (with or without motion artifacts). Additionally or alternatively, the one or more signals may include a signal from another type of sensor. As one example, the one or more signals can be processed to provide the user with information regarding their heart rate, the number of steps taken, cadence information, the intensity of the activity, calorie consumption, and/or the user's speed. The information can be provided in real time and/or provided after the user has completed the physical activity. In one embodiment, the additional information can be displayed to the user (e.g., on display 104 in FIG. 1).


Two examples of a PPG signal that includes motion artifacts and a filtered PPG signal for different activities are shown in FIGS. 4 and 5. FIG. 4 illustrates a PPG signal 400 that includes motion artifacts for a user wearing the electronic device 100 shown in FIG. 1 while standing and walking without any arm movement. In the illustrated embodiment, the filtered PPG signal 402 represents the heartbeats of the user. Between time T0 and T1 the user is standing still. Consequently, both the PPG signal 400 and the filtered PPG signal 402 are substantially flat during that time period.


The user is walking in place without any substantial arm movement between the time period T1 and T2. After time T1, the PPG signal 400 includes appreciable positive and negative amplitude peaks 404, 406. Similarly, the filtered PPG signal 402 includes appreciable positive and negative amplitude peaks 408, 410. Each walking step can cause a peak amplitude in the filtered PPG signal 402 that is larger than in the PPG signal 400. At time T2, the user stops walking and begins standing still again and the PPG signal 400 and the filtered PPG signal 402 are substantially flat.


One or more characteristics of the PPG signal 400 can be analyzed to determine if the user is standing or walking. In some embodiments, the peak-to-peak distances and/or the frequencies of the peak amplitudes in the PPG signal may correlate to a physical activity. Additionally or alternatively, the shape of the PPG signal over a given time period can be analyzed to determine the type of physical activity the user is performing (i.e., walking in this illustrated embodiment). The given time period can be any period of time (or multiple periods of time) that occur during the PPG signal. For example, the given time period can be the period between time T1 and time T2, or the given time period can be one or more subset time periods between time T1 and time T2. As one example, the period between time T3 and time T4 can be analyzed to determine the physical activity of the user. Additionally or alternatively, the period between time T0 and time T4 can be analyzed.


In some embodiments, the values of the peak amplitudes and/or the distances between positive peak amplitudes and negative peak amplitudes over a given time period can be considered when determining the physical activity of the user. Additionally or alternatively, characteristics of the OHRM signal not described herein can be analyzed to determine the physical activity performed by the user.



FIG. 5 depicts a second example of a PPG signal 500 that includes motion artifacts and a filtered PPG signal 502 for a user wearing the electronic device shown in FIG. 1 while standing and squatting. Once again, the filtered PPG signal 502 represents the heartbeats of the user. The user is standing between time T0 and time T1, squatting between time T1 and time T2, and standing again after time T2. Squatting can cause a reduction in intensity 504 in the PPG signal and the heart rate can increase 506 when the user is in the squat position. As described earlier, one or more characteristics of the PPG signal can be analyzed to determine that the user is standing and/or squatting.


The OHRM signal and motion artifacts shown in FIG. 4 differs from the OHRM signal and motion artifacts in FIG. 5. Thus, physical activities can have distinct OHRM signals and motion artifacts, allowing an OHRM signal to be used to identify a specific physical activity of a user. In some embodiments, a user may calibrate a wearable electronic device by storing OHRM signals for a variety of different activities the user wants the electronic device to be able to identify.


Referring now to FIG. 6, there is shown a flowchart of a method for calibrating a wearable electronic device to determine one or more activities of a user. Initially, the user performs the physical activity he or she wants an electronic device to be able to detect using an OHRM signal (block 600). An OHRM signal that is based on the type of physical activity that is being performed is then received at block 602. The OHRM signal includes one or more motion artifacts that is produced by the physical activity of the user. Optionally, a sensor signal from one or more other types of sensors (e.g., motion and/or position sensors) can also be received at block 602. For example, a signal from a gyroscope and/or an accelerometer can be received at block 602.


The OHRM signal having one or more motion artifacts and optionally a signal from one or more other sensors can be stored in memory. As one example, the OHRM signal can be stored in memory 202 or memory 224 shown in FIG. 2. Other embodiments can store data and/or characteristics of the OHRM signal and/or motion artifacts rather than the signal itself at block 604. Likewise, a signal and/or data and/or characteristics of the sensor signal that is received from one or more other sensors can be stored in the memory at block 604.


Next, as shown in block 606, the physical activity associated with the OHRM signal is identified and stored. An activity identifier can be received and associated with the OHRM signal. In one embodiment, the user can input an activity identifier using an input device included in the wearable electronic device. As one example, the user can input an activity identifier using a keyboard displayed on a touchscreen. In another embodiment, the user can speak the activity identifier and a voice recognition function can input the activity identification. In other embodiments, the activity identifier can be received from an external electronic device.


Subsequent OHRM signals that include one or more motion artifacts can then be correlated to an activity based on the associated activity identifier, and the identified activity may be displayed or provided to the user.


The method shown in FIG. 6 can be performed for each physical activity the user wants identified using an OHRM signal that includes motion artifacts. Other embodiments can perform the method differently. Some blocks can be omitted, new blocks added, and/or some of the blocks can be performed in a different order. For example, in some embodiments, an OHRM signal associated with a particular physical activity that is stored in memory can be updated or replaced with a newly captured OHRM signal associated with the same physical activity. A stored OHRM signal can be updated or replaced periodically, at select times, or each time the user performs the activity. In embodiments that store data and/or characteristics of the OHRM signal, the data and/or characteristics can be updated or replaced periodically, at select times, or each time the user performs the physical activity. As one example, a running average of a given OHRM signal can be maintained by updating the current OHRM signal for a particular activity with newly received OHRM signals that are determined to represent the same activity. As described previously, the OHRM signal or signals include one or more motion artifacts that is produced by the physical activity of the user.


Various embodiments have been described in detail with particular reference to certain features thereof, but it will be understood that variations and modifications can be effected within the spirit and scope of the disclosure. And even though specific embodiments have been described herein, it should be noted that the application is not limited to these embodiments. In particular, any features described with respect to one embodiment may also be used in other embodiments, where compatible. Likewise, the features of the different embodiments may be exchanged, where compatible.

Claims
  • 1. A method for determining a type of physical activity being performed by a user wearing an electronic device that includes a photoplethysmograph (PPG) sensor, the method comprising: receiving, by a processing device, an optical heart rate monitor (OHRM) signal from the PPG sensor;analyzing, by the processing device, the OHRM signal to detect one or more signal characteristics of the OHRM signal;analyzing, by the processing device, the one or more signal characteristics to detect repeated motion artifacts that are produced by the type of physical activity being performed by the user;based on the repeated motion artifacts: determining, by the processing device, the type of physical activity being performed by the user;identifying a time period in which the determined type of physical activity is being performed by the user; anddetermining, by the processing device and using the repeated motion artifacts during the identified time period, at least one of a number of steps taken by the user during the type of physical activity or a calorie consumption by the user during the type of physical activity; andproviding, to the user, information regarding the type of physical activity and at least one of the number of steps taken by the user during the type of physical activity or the calorie consumption by the user during the type of physical activity.
  • 2. The method as in claim 1, further comprising: receiving a sensor signal from at least one other type of sensor; andanalyzing the sensor signal when determining the type of physical activity being performed by the user.
  • 3. The method as in claim 1, wherein analyzing the OHRM signal comprises analyzing at least one amplitude peak in the OHRM signal during a given time period.
  • 4. The method as in claim 1, wherein analyzing the OHRM signal comprises analyzing a shape of the OHRM signal during a given time period.
  • 5. The method as in claim 1, wherein analyzing the OHRM signal comprises analyzing a frequency of the OHRM signal during a given time period.
  • 6. The method as in claim 1, wherein the electronic device comprises a device that provides time.
  • 7. The method as in claim 1, wherein the electronic device comprises a health assistant.
  • 8. An electronic device wearable by a user, comprising: a photoplethysmograph (PPG) sensor; andprocessing device that is operatively connected to the PPG sensor and adapted to: receive an optical heart rate monitor (OHRM) signal from the PPG sensor;analyze the OHRM signal to detect one or more signal characteristics of the OHRM signal;analyze the one or more signal characteristics to detect repeated motion artifacts that are produced by a type of physical activity being performed by a user;based on the repeated motion artifacts, determine the type of physical activity being performed by the user;identify a time period in which the determined type of physical activity is being performed by the user; anddetermine, using the repeated motion artifacts during the identified time period, at least one of a number of steps taken by the user during the type of physical activity or a calorie consumption by the user during the type of physical activity; andprovide, to the user, information regarding the type of physical activity and at least one of the number of steps or the calorie consumption.
  • 9. The electronic device as in claim 8, further comprising a memory adapted to store one or more OHRM signals that each represent a particular type of physical activity.
  • 10. The electronic device as in claim 8, wherein the processing device analyzes the OHRM signal by analyzing at least one amplitude peak in the OHRM signal during a given time period.
  • 11. The electronic device as in claim 8, wherein the processing device analyzes the OHRM signal by analyzing a shape of the OHRM signal during a given time period.
  • 12. The electronic device as in claim 8, wherein the processing device analyzes the OHRM signal by analyzing a frequency of the OHRM signal during a given time period.
  • 13. The electronic device as in claim 8, wherein the electronic device comprises a health assistant.
  • 14. The electronic device as in claim 8, wherein the electronic device comprises a device that provides time.
  • 15. A system, comprising: a photoplethysmograph (PPG) sensor;a motion or position sensor; anda processing device operatively connected to the PPG sensor and the motion or position sensor, wherein the processing device is adapted to: receive from the PPG sensor an OHRM signal;determine one or more signal characteristics of the OHRM signal;analyze the one or more signal characteristics to detect repeated motion artifacts that are produced by a type of physical activity being performed by a user;determine the type of physical activity being performed by the user based on the repeated motion artifacts;receive a sensor signal from the motion or position sensor;analyze the sensor signal to further categorize the type of physical activity being performed by the user;identify, based on the repeated motion artifacts, a time period in which the determined type of physical activity is being performed by the user;determine, using the repeated motion artifacts during the identified time period, at least one of a number of steps taken by the user during the type of physical activity or a calorie consumption by the user during the type of physical activity; andprovide, to the user, information regarding the type of physical activity and at least one of the number of steps or the calorie consumption.
  • 16. The system as in claim 15, wherein the system comprises a wearable electronic device.
  • 17. The system as in claim 16, wherein the PPG sensor and the processing device are included in a wearable electronic device and the motion or position sensor is included in an external electronic device communicably connected to the PPG sensor.
  • 18. A method for operating a wearable electronic device for determining a type of physical activity being performed by a user, the wearable electronic device including a photoplethysmograph (PPG) sensor, the method comprising: receiving, by a processing device, an optical heart rate monitor (OHRM) signal from the PPG sensor when a user performs a particular type of physical activity;determining first signal characteristics of the OHRM signal;detecting first motion artifacts that are produced by the particular type of physical activity based on the first signal characteristics;receiving an activity identifier from a user of the wearable electronic device;associating the activity identifier to the OHRM signal to associate the particular type of physical activity to the OHRM signal;storing, in a memory of the wearable electronic device, the activity identifier and the OHRM signal;determining a subsequently received a OHRM signal is associated with the particular type of physical activity, by: determining second signal characteristics of the subsequently received OHRM signal;detecting second motion artifacts based on the second signal characteristics; andcomparing the second motion artifacts with the first motion artifacts;identifying, based on the second motion artifacts, a time period in which the particular type of physical activity is being performed by the user;determining, using the second motion artifacts during the identified time period, at least one of a number of steps taken by the user during the particular type of physical activity or a calorie consumption by the user during the particular type of physical activity; andproviding, to the user, information regarding the type of physical activity associated with the subsequently received OHRM and at least one of the number of steps or the calorie consumption; andupdating the stored OHRM signal with the subsequently received OHRM signal, the updating comprising maintaining a running average of the OHRM signal associated with the particular type of physical activity.
CROSS-REFERENCE TO RELATED APPLICATION

This application claims the benefit under 35 U.S.C. § 119(e) of U.S. Provisional Patent Application No. 61/940,364, filed Feb. 14, 2014, entitled “Activity Identification Using An Optical Heart Rate Monitor,” the entirety of which is incorporated herein by reference.

US Referenced Citations (242)
Number Name Date Kind
4686572 Takatsu Aug 1987 A
4686648 Fossum Aug 1987 A
5105264 Erhardt et al. Apr 1992 A
5329313 Keith Jul 1994 A
5396893 Oberg et al. Mar 1995 A
5471515 Fossum et al. Nov 1995 A
5541402 Ackland Jul 1996 A
5550677 Schofield et al. Aug 1996 A
5781312 Noda Jul 1998 A
5841126 Fossum et al. Nov 1998 A
5880459 Pryor et al. Mar 1999 A
5949483 Fossum et al. Sep 1999 A
6008486 Stam et al. Dec 1999 A
6040568 Caulfield et al. Mar 2000 A
6233013 Hosier et al. May 2001 B1
6348929 Acharya et al. Feb 2002 B1
6448550 Nishimura Sep 2002 B1
6528833 Lee et al. Mar 2003 B2
6541751 Bidermann Apr 2003 B1
6670904 Yakovlev Dec 2003 B1
6713796 Fox Mar 2004 B1
6714239 Guidash Mar 2004 B2
6798453 Kaifu Sep 2004 B1
6816676 Bianchi et al. Nov 2004 B2
6905470 Lee et al. Jun 2005 B2
6956605 Hashimoto Oct 2005 B1
6982759 Goto Jan 2006 B2
7075049 Rhodes et al. Jul 2006 B2
7091466 Bock Aug 2006 B2
7119322 Hong Oct 2006 B2
7133073 Neter Nov 2006 B1
7259413 Rhodes Aug 2007 B2
7262401 Hopper et al. Aug 2007 B2
7271835 Iizuka et al. Sep 2007 B2
7282028 Kim et al. Oct 2007 B2
7319218 Krymski Jan 2008 B2
7332786 Altice Feb 2008 B2
7390687 Boettiger Jun 2008 B2
7415096 Sherman Aug 2008 B2
7437013 Anderson Oct 2008 B2
7443421 Stavely et al. Oct 2008 B2
7446812 Ando et al. Nov 2008 B2
7471315 Silsby et al. Dec 2008 B2
7502054 Kalapathy Mar 2009 B2
7525168 Hsieh Apr 2009 B2
7554067 Zarnoski et al. Jun 2009 B2
7555158 Park et al. Jun 2009 B2
7589316 Dunki-Jacobs Sep 2009 B2
7626626 Panicacci Oct 2009 B2
7622699 Sakakibara et al. Nov 2009 B2
7636109 Nakajima et al. Dec 2009 B2
7667400 Goushcha Feb 2010 B1
7671435 Ahn Mar 2010 B2
7714292 Agarwal et al. May 2010 B2
7728351 Shim Jun 2010 B2
7733402 Egawa et al. Jun 2010 B2
7742090 Street Jun 2010 B2
7764312 Ono et al. Jul 2010 B2
7773138 Lahav et al. Aug 2010 B2
7786543 Hsieh Aug 2010 B2
7796171 Gardner Sep 2010 B2
7817198 Kang et al. Oct 2010 B2
7838956 McCarten et al. Nov 2010 B2
7873236 Li et al. Jan 2011 B2
7880785 Gallagher Feb 2011 B2
7884402 Ki Feb 2011 B2
7906826 Martin et al. Mar 2011 B2
7952121 Arimoto May 2011 B2
7952635 Lauxtermann May 2011 B2
7982789 Watanabe et al. Jul 2011 B2
8026966 Altice Sep 2011 B2
8032206 Farazi et al. Oct 2011 B1
8089036 Manabe et al. Jan 2012 B2
8089524 Urisaka Jan 2012 B2
8094232 Kusaka Jan 2012 B2
8116540 Dean Feb 2012 B2
8140143 Picard et al. Mar 2012 B2
8153947 Barbier et al. Apr 2012 B2
8159570 Negishi Apr 2012 B2
8159588 Boemler Apr 2012 B2
8164669 Compton et al. Apr 2012 B2
8174595 Honda et al. May 2012 B2
8184188 Yaghmai May 2012 B2
8194148 Doida Jun 2012 B2
8194165 Border et al. Jun 2012 B2
8222586 Lee Jul 2012 B2
8227844 Adkisson et al. Jul 2012 B2
8233071 Takeda Jul 2012 B2
8259228 Wei et al. Sep 2012 B2
8310577 Neter Nov 2012 B1
8324553 Lee Dec 2012 B2
8338856 Tai et al. Dec 2012 B2
8340407 Kalman Dec 2012 B2
8350940 Smith et al. Jan 2013 B2
8355117 Niclass Jan 2013 B2
8388346 Rantala et al. Mar 2013 B2
8400546 Itano et al. Mar 2013 B2
8456540 Egawa Jun 2013 B2
8456559 Yamashita Jun 2013 B2
8508637 Han et al. Aug 2013 B2
8514308 Itonaga et al. Aug 2013 B2
8520913 Dean Aug 2013 B2
8546737 Tian et al. Oct 2013 B2
8547388 Cheng Oct 2013 B2
8575531 Hynecek et al. Nov 2013 B2
8581992 Hamada Nov 2013 B2
8594170 Mombers et al. Nov 2013 B2
8619163 Ogua Dec 2013 B2
8619170 Mabuchi Dec 2013 B2
8629484 Ohri et al. Jan 2014 B2
8634002 Kita Jan 2014 B2
8637875 Finkelstein et al. Jan 2014 B2
8648947 Sato et al. Feb 2014 B2
8653434 Johnson et al. Feb 2014 B2
8723975 Solhusvik May 2014 B2
8724096 Gosch et al. May 2014 B2
8730345 Watanabe May 2014 B2
8754983 Sutton Jun 2014 B2
8755854 Addison et al. Jun 2014 B2
8759736 Yoo Jun 2014 B2
8760413 Peterson et al. Jun 2014 B2
8767104 Makino et al. Jul 2014 B2
8803990 Smith Aug 2014 B2
8817154 Manabe et al. Aug 2014 B2
8879686 Okada et al. Nov 2014 B2
8902330 Theuwissen Dec 2014 B2
8908073 Minagawa Dec 2014 B2
8934030 Kim et al. Jan 2015 B2
8946610 Iwabuchi et al. Feb 2015 B2
8982237 Chen Mar 2015 B2
9041837 Li May 2015 B2
9017748 Theuwissen Jun 2015 B2
9054009 Oike et al. Jun 2015 B2
9058081 Baxter Jun 2015 B2
9066017 Geiss Jun 2015 B2
9066660 Watson et al. Jun 2015 B2
9088727 Trumbo Jul 2015 B2
9094623 Kawaguchi Jul 2015 B2
9099604 Roy Aug 2015 B2
9100597 Hu Aug 2015 B2
9106859 Kizuna et al. Aug 2015 B2
9131171 Aoki et al. Sep 2015 B2
9160949 Zhang et al. Oct 2015 B2
9164144 Dolinsky Oct 2015 B2
9178100 Webster et al. Nov 2015 B2
9209320 Webster Dec 2015 B1
9232150 Kleekajai et al. Jan 2016 B2
9232161 Suh Jan 2016 B2
9235267 Burrough et al. Jan 2016 B2
9270906 Peng et al. Feb 2016 B2
9287304 Park et al. Mar 2016 B2
9288380 Nomura Mar 2016 B2
9331116 Webster May 2016 B2
9344649 Bock May 2016 B2
9417326 Niclass et al. Aug 2016 B2
9438258 Yoo Sep 2016 B1
9445018 Fettig et al. Sep 2016 B2
9448110 Wong Sep 2016 B2
9478030 Lecky Oct 2016 B1
9497397 Kleekajai et al. Nov 2016 B1
9516244 Borowski Dec 2016 B2
9560339 Borowski Jan 2017 B2
9584743 Lin et al. Feb 2017 B1
9596423 Molgaard Mar 2017 B1
9749556 Fettig et al. Aug 2017 B2
9774318 Song Sep 2017 B2
9781368 Song Oct 2017 B2
9831283 Shepard et al. Nov 2017 B2
9888198 Mauritzson et al. Feb 2018 B2
9894304 Smith Feb 2018 B1
9912883 Agranov et al. Mar 2018 B1
10136090 Vogelsang et al. Nov 2018 B2
10153310 Zhang et al. Dec 2018 B2
20030036685 Goodman et al. Feb 2003 A1
20040207836 Chhibber et al. Oct 2004 A1
20050026332 Fratti et al. Feb 2005 A1
20050049470 Terry Mar 2005 A1
20060274161 Ing et al. Dec 2006 A1
20070263099 Motta et al. Nov 2007 A1
20080177162 Bae et al. Jul 2008 A1
20080315198 Jung Dec 2008 A1
20090096901 Bae et al. Apr 2009 A1
20090101914 Hirotsu et al. Apr 2009 A1
20090146234 Luo et al. Jun 2009 A1
20090201400 Zhang et al. Aug 2009 A1
20090219266 Lim et al. Sep 2009 A1
20100134631 Voth Jun 2010 A1
20110080500 Wang et al. Apr 2011 A1
20110152637 Kateraas Jun 2011 A1
20110156197 Tivarus et al. Jun 2011 A1
20110164162 Kato Jul 2011 A1
20110193824 Modarres et al. Aug 2011 A1
20110245690 Watson et al. Oct 2011 A1
20120092541 Tuulos et al. Apr 2012 A1
20120098964 Oggier et al. Apr 2012 A1
20120127088 Pance et al. May 2012 A1
20120147207 Itonaga Jun 2012 A1
20120239173 Laikari Sep 2012 A1
20130147981 Wu Jun 2013 A1
20130155271 Ishii Jun 2013 A1
20130222584 Aoki et al. Aug 2013 A1
20140049683 Guenter Feb 2014 A1
20140071321 Seyama Mar 2014 A1
20140132528 Catton May 2014 A1
20140167973 Letchner Jun 2014 A1
20140231630 Rae et al. Aug 2014 A1
20140240550 Taniguchi Aug 2014 A1
20140246568 Wan Sep 2014 A1
20140247378 Sharma et al. Sep 2014 A1
20140252201 Li et al. Sep 2014 A1
20140253754 Papiashvili Sep 2014 A1
20140263951 Fan et al. Sep 2014 A1
20140267855 Fan Sep 2014 A1
20140347533 Toyoda Nov 2014 A1
20140354861 Pang Dec 2014 A1
20150062391 Murata Mar 2015 A1
20150163392 Malone et al. Jun 2015 A1
20150163422 Fan et al. Jun 2015 A1
20150215443 Heo Jul 2015 A1
20150237314 Hasegawa Aug 2015 A1
20150264241 Kleekajai et al. Sep 2015 A1
20150264278 Kleekajai et al. Sep 2015 A1
20150277559 Vescovi et al. Oct 2015 A1
20150312479 McMahon et al. Oct 2015 A1
20150350575 Agranov et al. Dec 2015 A1
20160050379 Jiang et al. Feb 2016 A1
20160099371 Webster Apr 2016 A1
20160205311 Mandelli et al. Jul 2016 A1
20160218236 Dhulla et al. Jul 2016 A1
20160219232 Murata Jul 2016 A1
20160274237 Stutz Sep 2016 A1
20160307325 Wang et al. Oct 2016 A1
20160356890 Fried et al. Dec 2016 A1
20160365380 Wan Dec 2016 A1
20170047363 Choi et al. Feb 2017 A1
20170082746 Kubota et al. Mar 2017 A1
20170084133 Cardinali et al. Mar 2017 A1
20170142325 Shimokawa et al. May 2017 A1
20170223292 Ikeda Aug 2017 A1
20170272675 Kobayashi Sep 2017 A1
20170373106 Li et al. Dec 2017 A1
20180213205 Oh Jul 2018 A1
Foreign Referenced Citations (85)
Number Date Country
1630350 Jun 2005 CN
1774032 May 2006 CN
1833429 Sep 2006 CN
1842138 Oct 2006 CN
1947414 Apr 2007 CN
101189885 May 2008 CN
101221965 Jul 2008 CN
101233763 Jul 2008 CN
101472059 Jul 2009 CN
101567977 Oct 2009 CN
101622859 Jan 2010 CN
101739955 Jun 2010 CN
101754029 Jun 2010 CN
101803925 Aug 2010 CN
102036020 Apr 2011 CN
102067584 May 2011 CN
102208423 Oct 2011 CN
102451160 May 2012 CN
102668542 Sep 2012 CN
102820309 Dec 2012 CN
102821255 Dec 2012 CN
103024297 Apr 2013 CN
103051843 Apr 2013 CN
103329513 Sep 2013 CN
103546702 Jan 2014 CN
204761615 Nov 2015 CN
1763228 Mar 2007 EP
2023611 Feb 2009 EP
2107610 Oct 2009 EP
2230690 Sep 2010 EP
2512126 Oct 2012 EP
2787531 Oct 2014 EP
S61123287 Jun 1986 JP
2007504670 Aug 1987 JP
2000059697 Feb 2000 JP
2001211455 Aug 2001 JP
2001358994 Dec 2001 JP
2004111590 Apr 2004 JP
2005318504 Nov 2005 JP
2006287361 Oct 2006 JP
2007516654 Jun 2007 JP
2008507908 Mar 2008 JP
2008271280 Nov 2008 JP
2008543061 Nov 2008 JP
2009021809 Jan 2009 JP
2009159186 Jul 2009 JP
2009212909 Sep 2009 JP
2009296465 Dec 2009 JP
2010080604 Apr 2010 JP
2010114834 May 2010 JP
2011040926 Feb 2011 JP
201149697 Mar 2011 JP
2011091775 May 2011 JP
2011097646 Dec 2011 JP
2012010306 Jan 2012 JP
2012019516 Jan 2012 JP
2012513160 Jun 2012 JP
2013051523 Mar 2013 JP
2013070240 Apr 2013 JP
2013529035 Jul 2013 JP
20030034424 May 2003 KR
20030061157 Jul 2003 KR
20050103732 Nov 2005 KR
20080069851 Jul 2008 KR
20100008239 Jan 2010 KR
20100065084 Jun 2010 KR
20130074459 Jul 2013 KR
200520551 Jun 2005 TW
200803481 Jan 2008 TW
201110689 Mar 2011 TW
201301881 Jan 2013 TW
WO 05041304 May 2005 WO
WO 06014641 Feb 2006 WO
WO 06130443 Dec 2006 WO
WO 07049900 May 2007 WO
WO 10120945 Oct 2010 WO
WO 12011095 Jan 2012 WO
WO 12032353 Mar 2012 WO
WO 12053363 Apr 2012 WO
WO 12088338 Jun 2012 WO
WO 12122572 Sep 2012 WO
WO 12138687 Oct 2012 WO
WO 13008425 Jan 2013 WO
WO 13179018 Dec 2013 WO
WO 13179020 Dec 2013 WO
Non-Patent Literature Citations (50)
Entry
Aoki, et al., “Rolling-Shutter Distortion-Free 3D Stacked Image Sensor with −160dB Parasitic Light Sensitivity In-Pixel Storage Node,” ISSCC 2013, Session 27, Image Sensors, 27.3 27.3 A, Feb. 20, 2013, retrieved on Apr. 11, 2014 from URL:http://ieeexplore.ieee.org/stamp/stamp.jsp?tp=&arnumber=6487824.
Elgendi, “On the Analysis of Fingertip Photoplethysmogram Signals,” Current Cardiology Reviews, 2012, vol. 8, pp. 14-25.
Feng, et al., “On the Stoney Formula for a Thin Film/Substrate System with Nonuniform Substrate Thickness,” Journal of Applied Mechanics, Transactions of the ASME, vol. 74, Nov. 2007, pp. 1276-1281.
Fu, et al., “Heart Rate Extraction from Photoplethysmogram Waveform Using Wavelet Multui-resolution Analysis,” Journal of Medical and Biological Engineering, 2008, vol. 28, No. 4, pp. 229-232.
Han, et al., “Artifacts in wearable photoplethysmographs during daily life motions and their reduction with least mean square based active noise cancellation method,” Computers in Biology and Medicine, 2012, vol. 42, pp. 387-393.
Lopez-Silva, et al., “Heuristic Algorithm for Photoplethysmographic Heart Rate Tracking During Maximal Exercise Test,” Journal of Medical and Biological Engineering, 2011, vol. 12, No. 3, pp. 181-188.
Santos, et al., “Accelerometer-assisted PPG Measurement During Physical Exercise Using the LAVIMO Sensor System,” Acta Polytechnica, 2012, vol. 52, No. 5, pp. 80-85.
Sarkar, et al., “Fingertip Pulse Wave (PPG signal) Analysis and Heart Rate Detection,” International Journal of Emerging Technology and Advanced Engineering, 2012, vol. 2, No. 9, pp. 404-407.
Schwarzer, et al., On the determination of film stress from substrate bending: Stoney's formula and its limits, Jan. 2006, 19 pages.
Yan, et al., “Reduction of motion artifact in pulse oximetry by smoothed pseudo Wigner-Ville distribution,” Journal of NeuroEngineering and Rehabilitation, 2005, vol. 2, No. 3, pp. 1-9.
Yousefi, et al., “Adaptive Cancellation of Motion Artifact in Wearable Biosensors,” 34th Annual International Conference of the IEEE EMBS, San Diego, California, Aug./Sep. 2012, pp. 2004-2008.
U.S. Appl. No. 15/056,752, filed Feb. 29, 2016, Wan.
U.S. Appl. No. 15/590,775, filed May 9, 2017, Lee.
Shen et al., “Stresses, Curvatures, and Shape Changes Arising from Patterned Lines on Silicon Wafers,” Journal of Applied Physics, vol. 80, No. 3, Aug. 1996, pp. 1388-1398.
U.S. Appl. No. 15/627,409, filed Jun. 19, 2017, Agranov et al.
U.S. Appl. No. 15/653,458, filed Jul. 18, 2017, Zhang et al.
U.S. Appl. No. 15/682,255, filed Aug. 21, 2017, Li et al.
U.S. Appl. No. 15/699,806, filed Sep. 8, 2017, Li et al.
U.S. Appl. No. 15/713,477, filed Sep. 22, 2017, Mandai et al.
U.S. Appl. No. 15/713,520, filed Sep. 22, 2017, Mandai et al.
Charbon, et al., SPAD-Based Sensors, TOF Range-Imaging Cameras, F. Remondino and D. Stoppa (eds.), 2013, Springer-Verlag Berlin Heidelberg, pp. 11-38.
Cox, “Getting histograms with varying bin widths,” http://www.stata.com/support/faqs/graphics/histograms-with-varying-bin-widths/, Nov. 13, 2017, 5 pages.
Gallivanoni, et al., “Progress n Quenching Circuits for Single Photon Avalanche Diodes,” IEEE Transactions on Nuclear Science, vol. 57, No. 6, Dec. 2010, pp. 3815-3826.
Leslar, et al., “Comprehensive Utilization of Temporal and Spatial Domain Outlier Detection Methods for Mobile Terrestrial LiDAR Data,” Remote Sensing, 2011, vol. 3, pp. 1724-1742.
Mota, et al., “A flexible multi-channel high-resolution Time-to-Digital Converter ASIC,” Nuclear Science Symposium Conference Record IEEE, 2000, Engineering School of Geneva, Microelectronics Lab, Geneva, Switzerland, 8 pages.
Niclass, et al., “Design and Characterization of a CMOS 3-D Image Sensor Based on Single Photon Avalanche Diodes,” IEEE Journal of Solid-State Circuits, vol. 40, No. 9, Sep. 2005, pp. 1847-1854.
Shin, et al., “Photon-Efficient Computational 3D and Reflectivity Imaging with Single-Photon Detectors,” IEEE International Conference on Image Processing, Paris, France, Oct. 2014, 11 pages.
Tisa, et al., “Variable-Load Quenching Circuit for single-photon avalanche diodes,” Optics Express, vol. 16, No. 3, Feb. 4, 2008, pp. 2232-2244.
Ullrich, et al., “Linear LIDAR versus Geiger-mode LIDAR: Impact on data properties and data quality,” Laser Radar Technology and Applications XXI, edited by Monte D. Turner, Gary W. Kamerman, Proc. of SPIE, vol. 9832, 983204, 2016, 17 pages.
U.S. Appl. No. 15/879,365, filed Jan. 24, 2018, Mandai et al.
U.S. Appl. No. 15/879,350, filed Jan. 24, 2018, Mandai et al.
U.S. Appl. No. 15/880,285, filed Jan. 25, 2018, Laifenfeld et al.
U.S. Appl. No. 13/782,532, filed Mar. 1, 2013, Sharma et al.
U.S. Appl. No. 13/783,536, filed Mar. 4, 2013, Wan.
U.S. Appl. No. 13/785,070, filed Mar. 5, 2013, Li.
U.S. Appl. No. 13/787,094, filed Mar. 6, 2013, Li et al.
U.S. Appl. No. 13/797,851, filed Mar. 12, 2013, Li.
U.S. Appl. No. 13/830,748, filed Mar. 14, 2013, Fan.
U.S. Appl. No. 14/098,504, filed Dec. 5, 2013, Fan et al.
U.S. Appl. No. 14/207,150, filed Mar. 12, 2014, Kleekajai et al.
U.S. Appl. No. 14/207,176, filed Mar. 12, 2014, Kleekajai et al.
U.S. Appl. No. 14/276,728, filed May 13, 2014, McMahon et al.
U.S. Appl. No. 14/292,599, filed May 30, 2014, Agranov et al.
U.S. Appl. No. 14/462,032, filed Aug. 18, 2014, Jiang et al.
U.S. Appl. No. 14/481,806, filed Sep. 9, 2014, Kleekajai et al.
U.S. Appl. No. 14/481,820, filed Sep. 9, 2014, Lin et al.
U.S. Appl. No. 14/501,429, filed Sep. 30, 2014, Malone et al.
U.S. Appl. No. 14/503,322, filed Sep. 30, 2014, Molgaard.
U.S. Appl. No. 14/611,917, filed Feb. 2, 2015, Lee et al.
Jahromi et al., “A Single Chip Laser Radar Receiver with a 9x9 SPAD Detector Array and a 10-channel TDC,” 2013 Proceedings of the ESSCIRC, IEEE, Sep. 14, 2015, pp. 364-367.
Provisional Applications (1)
Number Date Country
61940364 Feb 2014 US