Methods, systems and devices for measuring heart rate

Information

  • Patent Grant
  • 11259707
  • Patent Number
    11,259,707
  • Date Filed
    Thursday, November 30, 2017
    6 years ago
  • Date Issued
    Tuesday, March 1, 2022
    2 years ago
Abstract
A system and method of tracking activity includes a motion sensor, a light source and a light detector. The light detector is configured to capture an amount of the light that is reflected back to the light detector, at least a first portion of the light reflected back to the light detector is reflected from a blood vessel disposed under a skin adjacent to the housing. A processor is in communication with the motion sensor and the light detector and can process the reflected light to identify heart beats and produce an indication of a heart rate. The indication of the heart rate can be displayed on the display screen as an option, in addition to the metrics that quantify the motion data.
Description
BACKGROUND

The present disclosure relates to systems and methods for capturing bodily activity and synchronizing data transfers between a capture device and a client device.


In recent years, the need for health and fitness has grown tremendously. The growth has occurred due to a better understanding of the benefits of good fitness to overall health and wellness. Unfortunately, although today's modern culture has brought about many new technologies, such as the Internet, connected devices and computers, people have become less active. Additionally, many office jobs require people to sit in front of computer screens for long periods of time, which further reduces a person's activity levels. Furthermore, much of today's entertainment options involve viewing multimedia content, computer social networking, and other types of computer involved interfacing. Although such computer activity can be very productive as well as entertaining, such activity tends to reduce a person's overall physical activity.


To provide users concerned with health and fitness a way of measuring or accounting for their activity or lack thereof, fitness trackers are often used. Fitness trackers are used to measure activity, such as walking, motion, running, sleeping, being inactive, bicycling, exercising on an elliptical trainer, and the like. Usually, the data collected by such fitness trackers can be transferred and viewed on a computing device. However, such data is often provided as a basic accumulation of activity data with complicated or confusing interfaces. In addition, updates between a tracker and a client device usually require wired connectors and/or complex syncing schemes.


It is in this context that embodiments described herein arise.


SUMMARY

Broadly speaking, the present invention fills these needs by providing a system and method for detecting and measuring a user's heart rate. It should be appreciated that the present invention can be implemented in numerous ways, including as a process, an apparatus, a system, computer readable media, or a device. Several inventive embodiments of the present invention are described below.


One embodiment provides an activity tracking device having a motion sensor and a processor in a housing. The processor configured for processing motion data produced by the motion sensor. A display screen is integrated with the housing to display metrics that quantify the motion data produced by the motion sensor. A light source is integrated within the housing to enable light to be directed out of the housing at a heart rate monitor location on the housing and a light detector is integrated within the housing. The light detector configured to capture an amount of the light that is reflected back to the light detector, at least a first portion of the light reflected back to the light detector is reflected from a blood vessel under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor is also in communication with the light detector to process the reflected light to identify heart beats of the user and produce an indication of a heart rate that can be displayed on the display screen, as an option, in addition to the metrics that quantify the motion data.


The processor can differentiate between a baseline light scattering and reflectance signal detected between the each one of multiple heart beats and a second light scattering and reflectance signal corresponding to at least one heart beat in the blood vessel. The second reflectance signal being less than the baseline reflectance signal, where the blood vessel scatters more of the light during the at least one heart beat than between each one of the multiple heart beats.


The motion sensor can be one of or include an accelerometer, or a global positioning sensor, or a magnetometer, or a gyroscope, or a rotary encoder, or a calorie measurement sensor, or a moisture measurement sensor, or a displacement sensor, or an ultrasonic sensor, or a pedometer, or an altimeter, or a linear motion sensor, or an angular motion sensor, or a multi-axis motion sensor, or a combination of two or more thereof.


The activity tracking device can also include a communication transceiver configured for communicating via at least one a wireless network, an ambient light sensor, an indicator for visually identifying the heart rate monitor location on the housing and at least one infrared (IR) proximity sensor associated with the light source and light detector. The IR proximity sensor can be configured to activate the light source and light detector upon detecting presence of the skin of the user. Detecting presence of the skin of the user can also function to navigate to one or more metrics of the display screen.


The activity tracking device can also include a pressure detecting system configured for detecting a pressure applied to the heart rate monitor location on the housing with the skin of the user during the identification of heart beats. The activity tracking device can output at least one a feedback signal regarding the detected pressure applied to the heart rate monitor location, the feedback indication being indicative of more or less pressure desired to produce the heart rate, the feedback signal including at least one of a visual signal, a graphic signal, a tactile signal, and an audible signal. The pressure detecting system can include at least one of processing of the reflected light to identify one of an excess pressure, an insufficient pressure or an acceptable pressure from the detected heart beats of the user, or a pressure sensor in the button.


The activity tracking device can also include a display of one or both of waveform data or numerical data when the skin of the user is over the heart rate monitor location and the heart beats are being identified over a sampling time period and upon concluding the sampling time period, displaying the heart rate on the display screen. At least one recalled heart rate can be displayed on the display screen. The light detector can be disposed next to the light source.


The heart rate can be calculated based on an algorithm that detects multiple heart beats in the light received in the light detector within a sampling time period, measures a first time interval between a first beat of the detected heart beats and a second beat of the detected heart beats and divides the sample time interval by the first time interval to determine a first estimate of heart beats detected within the sampling time period. The first estimate of heart beats is extrapolated within the sampling time period to a first estimated heart beats per minute and the first estimate heart beats per minute is output to the display screen.


The heart rate can be calculated based on an algorithm that adds at least one beat to the first estimate of heart beats to produce a second estimate of heart beats and subtracts at least one beat from the first estimate of beats to produce a third estimate of heart beats. The first estimate of heart beats, the second estimate of heart beats and the third estimate of heart beats are scored and a highest scoring estimate of heart beats is selected and output to the display screen.


The light source and the reflected light detector can be selected for any suitable wavelength or suitable band of wavelengths of light ranging from between infrared wavelengths through a human visible spectrum to ultraviolet wavelengths. The light source includes at least one of an infrared (IR) light source, wherein the IR light emitted from the light source produces a deadfront at the heart rate monitor location of the housing or a green light source and the heart rate monitor location includes a translucent green window.


The heart rate monitor location includes a cover that enables infrared (IR) light of the light source or any other wavelength of light from the light source to pass while blocking substantially all light in a human visible spectrum. The light source and light detector can additionally function as a proximity sensor to activate the display screen. The heart rate monitor location can include a button. The button can have an infrared (IR) light transmitting structure. The light source and the light detector can be disposed substantially below the button. The button can also function to navigate to one or more metrics of the display screen. The functions to navigate can be enabled while the heart beats are measured. The skin can be of a finger of a user.


Another embodiment provides a method of tracking activity including sensing motion with a motion sensor, the motion sensor including a processor. The processor is configured for processing motion data produced by the motion sensor. Metrics that quantify the motion data produced by the motion sensing can be displayed on a device display screen integrated with a housing of the motion sensor. A directed light is emitted from a light source, the light source being included in the housing at a heart rate monitor location on the housing and an amount of the light that is reflected back to a light detector is captured by the light detector integrated within the housing. At least a first portion of the light reflected back to the light detector is reflected from a blood vessel disposed under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor further being in communication with the light detector to enable processing of the reflected light to identify heart beats of the user and produce an indication of a heart rate. The indication of the heart rate being displayable on the display screen as an option, in addition to the metrics that quantify the motion data.


The method can also include detecting a pressure applied to the heart rate monitor location of the housing with the skin of the user during the identification of heart beats, the pressure being detected by a pressure detecting system included in the housing and outputting a feedback signal regarding the detected pressure applied to the heart rate monitor location, the feedback indication being indicative of more or less pressure desired to produce the heart rate. One or both of waveform data or numerical data can be displayed when the skin of the user is over the heart rate monitor location and the heart beats are being identified over a sampling time period. The heart rate can be displayed on the display screen upon concluding the sampling time period.


Yet another embodiment provides a heart rate monitor including a light source and a light detector disposed to receive light emitted from the light source and reflected from a blood vessel disposed within in a heart rate monitor subject. A display screen and a processor coupled to the light source, the light detector and the display screen, are also included.


Still another embodiment provides an activity tracking device including a housing including a motion sensor and a processor. The processor is configured for processing motion data produced by the motion sensor. A display screen is integrated with the housing to display metrics that quantify the motion data produced by the motion sensor. A light source is also integrated within the housing to enable light to be directed out of the housing at a heart rate monitor location on the housing and a light detector is integrated within the housing. The light detector is configured to capture an amount of the light that is reflected back to the light detector. At least a first portion of the light reflected back to the light detector is reflected from a blood vessel disposed under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor is also in communication with the light detector to enable processing of the reflected light to identify heart beats of the user and produce an indication of a heart rate that can be displayed on the display screen, as an option, in addition to the metrics that quantify the motion data. Other aspects and advantages of the invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrating by way of example the principles of the invention.





BRIEF DESCRIPTION OF THE DRAWINGS

The present invention will be readily understood by the following detailed description in conjunction with the accompanying drawings.



FIG. 1A shows a block diagram of an activity tracking device, in accordance with embodiments of the present invention



FIG. 1B illustrates an example of an activity tracking device having a housing in the form of a wearable wrist attachable device.



FIG. 1C illustrates another example of an activity tracking device, in accordance with embodiments of the present invention.



FIG. 2A illustrates an example of activity tracking device of FIG. 1A, showing some additional example components utilized for tracking activity and motion of the device, and associated interfaces to display screen.



FIG. 2B illustrates an example of activity tracking device in communication with a remote device.



FIGS. 3A and 3B illustrate examples of activity tracking devices having a heart rate measuring system in the form of a wearable wrist attachable device, in accordance with embodiments of the present invention.



FIGS. 3C and 3D illustrate another example of an activity tracking device, in accordance with embodiments of the present invention.



FIG. 4A is a flowchart diagram of heart rate measuring system, in accordance with embodiments of the present invention.



FIG. 4B is a graphical representation of an example series of peaks corresponding to detected heart beats captured during a sampling time period STP, in accordance with embodiments of the present invention.



FIG. 4C is a graphical representation of the filtering process for the above example series of peaks, in accordance with embodiments of the present invention.



FIG. 5 is a flowchart diagram of the method operations for detecting a heart rate, in accordance with embodiments of the present invention.



FIGS. 6A-D illustrate different presentations of the calculated heart rate on the display screen, in accordance with embodiments of the present invention.



FIG. 7A is a flowchart diagram of the method operations for adjusting the user pressure on the activity tracking device while detecting a heart rate, in accordance with embodiments of the present invention.



FIGS. 7B and 7C show feedback signals on the activity tracing device display, in accordance with embodiments of the present invention.



FIG. 8 illustrates an example where various types of activities of users can be captured by activity tracking devices, in accordance with embodiments of the present invention.





DETAILED DESCRIPTION

Several exemplary embodiments for activity tracking devices and methods capable monitoring and displaying both a user's activities and the user's heart rate will now be described. It will be apparent to those skilled in the art that the present invention may be practiced without some or all of the specific details set forth herein.


The user's heart rate can be measured by directing a light of a suitable wavelength into a user's skin and capturing a portion of the light reflected from a user's blood vessel inside the user's body. The reflected light includes data corresponding to the user's heart beats. Using this heart beat data and filtering methods and systems provide a rapid, accurate measurement of the user's heart rate. Thereby allowing the user to monitor both his activity and his heart rate. Motion data from a motion sensor within the activity tracking device can be used to identify false heart beats and provide a more accurate heart rate indication to the user, even while the user is engaged in a rigorous activity.


Embodiments described in the present disclosure provide systems, apparatus, computer readable media, and methods for analyzing tracked activity data and providing navigation screens and interfaces. Some embodiments are directed to providing navigation interfaces for an activity tracking device. The activity tracking device includes sensors for detecting when physical contact occurs onto the activity tracking device and logic for providing a display action to the screen of the activity tracking device. The physical contact, in one embodiment, can be qualified as an input when the physical contact has a particular characteristic that is predefined. The characteristic can be, when the contact is the result of one or more taps, e.g., physical contact to the activity tracking device by a finger or hand of the user, or object held by a user and used to impart the contact.


In other embodiments, the input can be non-physical, such as proximity sensing input. The proximity sensing input can be processed by an infrared proximity sensor, a thermal sensor, etc. The input can also be by way of a button, voice input, gaze detected input, input processed in response to motion or motion profiles, etc.


It should be noted that there are many inventions described and illustrated herein. The present inventions are neither limited to any single aspect nor embodiment thereof, nor to any combinations and/or permutations of such aspects and/or embodiments. Moreover, each of the aspects of the present inventions, and/or embodiments thereof, may be employed alone or in combination with one or more of the other aspects of the present inventions and/or embodiments thereof. For the sake of brevity, many of those permutations and combinations will not be discussed separately herein.


Further, in the course of describing and illustrating the present inventions, various circuitry, architectures, structures, components, functions and/or elements, as well as combinations and/or permutations thereof, are set forth. It should be understood that circuitry, architectures, structures, components, functions and/or elements other than those specifically described and illustrated, are contemplated and are within the scope of the present inventions, as well as combinations and/or permutations thereof.



FIG. 1A shows a block diagram of an activity tracking device 100, in accordance with embodiments of the present invention. The activity tracking device 100 is contained in a housing 101, which may be worn or held by a user. The housing 101 may be in the form of a wristband, a clip on device, a wearable device, or may be held by the user either in the user's hand or in a pocket or attached to the user's body. The activity tracking device 100 includes device components 102, which may be in the form of logic, storage, and glue logic, one or more processors, microelectronics, and interfacing circuitry. In one example, the components 102 will include a processor 106, memory 108, a wireless transceiver 110, a user interface 114, biometric sensors 116, and environmental sensors 118.


The environmental sensors 118 may be in the form of motion detecting sensors 118A. In some embodiments, a motion sensor 118A can be one or more of an accelerometer, or a gyroscope, or a rotary encoder, or a calorie measurement sensor, or a heat measurement sensor, or a moisture measurement sensor, or a displacement sensor, or an ultrasonic sensor, or a pedometer, or an altimeter, or a linear motion sensor, or an angular motion sensor, or a multi-axis motion sensor, or a combination thereof.


The biometric sensors 116 can be defined to measure physiological characteristics of the user that is using the activity tracking device 100. The user interface 114 provides a way for communicating with the activity tracking device 100, in response to user interaction 104. The user interaction 104 can be in the form of physical contact (e.g., without limitation, tapping, sliding, rubbing, multiple taps, gestures, etc.). The biometric sensors 116 can be a one or more proximity sensors 184 capable of detecting the user's presence or touch within a predefined distance or proximity. The proximity sensor 184 can be an infrared (IR) proximity sensor associated with the light source 181 and light detector 182, the IR proximity sensor configured to activate the light source and light detector upon detecting presence of the skin of the user.


The light source 181 and the light detector 182 are located near the external surface of the activity tracking device 100 at a heart rate monitor location 183. The heart rate monitor location 183 can include an indicator such as a marking or an image so the user can easily identify the heart rate monitor location 183. The marking or image can be a raised dot or dimple or a depression or an image of the fingerprint or the heart or any other suitable indication of the heart rate monitor location 183. The heart rate monitor location 183 can include a cover that enables infrared (IR) light of the light source 181 to pass while blocking substantially all light in a human visible spectrum. The heart rate monitor location 183 can include the button 126 or be separate from the button. In one embodiment, the button 126 has an infrared (IR) light transmitting structure and the light source 181 and the light detector 182 are disposed below the button, inside the housing. The button 126 can also provide navigation functions to one or more metrics of the display screen 122.


In some embodiments, the user interface 114 is configured to receive user interaction 104 that is in the form of noncontact input. The noncontact input can be by way of one or more proximity sensors 184, button presses, touch sensitive screen inputs, graphical user interface inputs, voice inputs, sound inputs, etc. The activity tracking device 100 can communicate with a client and/or server 112 using the wireless transceiver 110. The wireless transceiver 110 will allow the activity tracking device 100 to communicate using a wireless connection, which is enabled by wireless communication logic. The wireless communication logic can be in the form of a circuit having radio communication capabilities. The radio communication capabilities can be in the form of a Wi-Fi connection, a Bluetooth connection, a low-energy Bluetooth connection, or any other form of wireless tethering or near field communication. In still other embodiments, the activity tracking device 100 can communicate with other computing devices using a wired connection (not shown). As mentioned, the environmental sensors 118 can detect motion of the activity tracking device 100.


The motion can be activity of the user, such as walking, running, stair climbing, etc. The motion can also be in the form of physical contact received on any surface of the activity tracking device 110, so long as the environmental sensors 118 can detect such motion from the physical contact. As will be explained in more detail below, the physical contact may be in the form of a tap or multiple taps by a finger upon the housing of the activity tracking device 100.



FIG. 1B illustrates an example of an activity tracking device 100 having a housing 130 in the form of a wearable wrist attachable device. The sensors of the activity tracking device 100 can, as mentioned above, detect motion such as physical contact that is applied and received on a surface 120 of the housing 130. In the example shown, the physical contact 124 is in the form of a tap or multiple taps on the surface 120. Device components 102 are, in one embodiment, contained within the housing 130. The location at which the device components 102 are integrated into the housing 130 can vary. For example, the device components 102 can be integrated throughout various locations around the housing 130, and not limited to the central portion of the wrist attachable device. In some embodiments, the device components 102 can be integrated into or with a smart watch device.


In other embodiments, the device components 102 are positioned substantially in a central position of the wrist attachable device, such as under or proximate to a location where a display screen 122 is located. In the illustrated example, the housing 130 also includes a button 126. The button 126 can be pressed to activate the display screen 122, navigate to various metrics displayed on the screen 122, or turn off the screen 122.



FIG. 1C illustrates another example of an activity tracking device 100, in accordance with embodiments of the present invention. The form factor of the activity tracking device 100 is shown as a clickable device that includes a screen 122, a button 126, and device components 102 integrated within the housing 130′. The housing 130′ can include a clip that allows for attachment to clothing or articles of the user, or to simply place the device within a pocket or holder of the user. Accordingly, the physical contact 124 such as a touch or a tap, as shown with respect to FIG. 1B, can also be implemented upon the surface 120 of activity tracking device 100 of FIG. 1C. It should be understood, therefore, that the form factor of the activity tracking device 100 can take on various configurations and should not be limited to the example configurations provided herein.



FIG. 2A illustrates an example of activity tracking device 100 of FIG. 1A, showing some additional example components utilized for tracking activity and motion of the device, and associated interfaces to display screen 122. In this example, the finger of a user can be used to tap and provide physical contact 124 onto any surface 120 of activity tracking device 100. The physical contact, when sensed by sensors 184 of the activity tracking device 100, will cause a response by the activity tracking device 100, and therefore provide some metric on the display screen 122. In one embodiment, examples of a display screen 122 can include, but are not limited to, liquid crystal display (LCD) screens, light emitting diode (LED) screens, organic light emitting diode (OLED) screens, plasma display screens, etc.


As shown in FIG. 2A, the activity tracking device 100 includes logic 158. Logic 158 may include activity tracking logic 140, physical contact logic 142, display interface logic 144, alarm management logic 146, wireless communication logic 148, processor 106, and sensors 184. Additionally, storage (e.g. memory) 108, and a battery 154 can be integrated within the activity tracking device 100. The activity tracking logic 140 can include logic that is configured to process motion data produced by motion sensors 118, so as to quantify the motion and produce identifiable metrics associated with the motion.


Some motions will produce and quantify various types of metrics, such as step count, stairs climbed, distance traveled, very active minutes, calories burned, etc. The physical contact logic 142 can include logic that calculates or determines when particular physical contact can qualify as an input. To qualify as an input, the physical contact detected by biometric sensors 116 should have a particular pattern that is identifiable as input. For example, the input may be predefined to be a double tap input, and the physical contact logic 142 can analyze the motion to determine if a double tap indeed occurred in response to analyzing the sensor data produced by sensors 116, 118.


In other embodiments, the physical contact logic can be programmed to determine when particular physical contacts occurred, the time in between the physical contacts, and whether the one or more physical contacts will qualify within predefined motion profiles that would indicate that an input is desired. If physical contact occurs that is not within some predefined profile or pattern, the physical contact logic will not indicate or qualify that physical contact as an input.


The display interface logic 144 is configured to interface with the processor and the physical contact logic to determine when specific metric data will be displayed on the display screen 122 of the activity tracking device 100. The display interface logic 144 can act to turn on the screen, display metric information, display characters or alphanumeric information, display graphical user interface graphics, or combinations thereof. Alarm management logic 146 can function to provide a user interface and settings for managing and receiving input from a user to set an alarm. The alarm management logic can interface with a timekeeping module (e.g., clock, calendar, time zone, etc.), and can trigger the activation of an alarm. The alarm can be in the form of an audible alarm or a non-audible alarm.


A non-audible alarm can provide such alarm by way of a vibration. The vibration can be produced by a motor integrated in the activity tracking device 100. The vibration can be defined to include various vibration patterns, intensities, and custom set patterns. The vibration produced by the motor or motors of the activity tracking device 100 can be managed by the alarm management logic 146 in conjunction with processing by the processor 106. The wireless communication logic 148 is configured for communication of the activity tracking device with another computing device by way of a wireless signal. The wireless signal can be in the form of a radio signal. As noted above, the radio signal can be in the form of a Wi-Fi signal, a Bluetooth signal, a low energy Bluetooth signal, or combinations thereof. The wireless communication logic can interface with the processor 106, storage 108 and battery 154 of device 100, for transferring activity data, which may be in the form of motion data or processed motion data, stored in the storage 108 to the computing device.


In one embodiment, processor 106 functions in conjunction with the various logic components 140, 142, 144, 146, and 148. The processor 106 can, in one embodiment, provide the functionality of any one or all of the logic components. In other embodiments, multiple chips can be used to separate the processing performed by any one of the logic components and the processor 106. Sensors 116, 118 can communicate via a bus with the processor 106 and/or the logic components. The storage 108 is also in communication with the bus for providing storage of the motion data processed or tracked by the activity tracking device 100. Battery 154 is provided for providing power to the activity tracking device 100.



FIG. 2B illustrates an example of activity tracking device 100 in communication with a remote device 200. Remote device 200 is a computing device that is capable of communicating wirelessly with activity tracking device 100 and with the Internet 160. Remote device 200 can support installation and execution of applications. Such applications can include an activity tracking application 202. Activity tracking application 202 can be downloaded from a server. The server 220 can be a specialized server or a server that provides applications to devices, such as an application store. Once the activity tracking application 202 is installed in the remote device 200, the remote device 200 can communicate or be set to communicate with activity tracking device 100 (Device A). The remote device 200 can be a smartphone, a handheld computer, a tablet computer, a laptop computer, a desktop computer, or any other computing device capable of wirelessly interfacing with Device A 100 and the Internet 160.


In one embodiment, remote device 200 communicates with activity tracking device 100 over a Bluetooth connection. In one embodiment, the Bluetooth connection is a low energy Bluetooth connection (e.g., Bluetooth LE, BLE, or Bluetooth Smart). Low energy Bluetooth is configured for providing low power consumption relative to standard Bluetooth circuitry. Low energy Bluetooth uses, in one embodiment, a 2.4 GHz radio frequency, which allows for dual mode devices to share a single radio antenna. In one embodiment, low energy Bluetooth connections can function at distances up to 50 meters, with over the air data rates ranging between 1-3 megabits (Mb) per second. In one embodiment, a proximity distance for communication can be defined by the particular wireless link, and is not tied to any specific standard. It should be understood that the proximity distance limitation will change in accordance with changes to existing standards and in view of future standards and/or circuitry and capabilities.


Remote device 200 can also communicate with the Internet 160 using an Internet connection. The Internet connection of the remote device 200 can include cellular connections, wireless connections such as Wi-Fi, and combinations thereof (such as connections to switches between different types of connection links). The remote device, as mentioned above, can be a smartphone or tablet computer, or any other type of computing device having access to the Internet and with capabilities for communicating with the activity tracking device 100.


A server 220 is also provided, which is interfaced with the Internet 160. The server 220 can include a number of applications that service the activity tracking device 100, and the associated users of the activity tracking device 100 by way of user accounts. For example, the server 220 can include an activity management application 224. The activity management application 224 can include logic for providing access to various devices 100, which are associated with user accounts managed by server 220. Server 220 can include storage 226 that includes various user profiles associated with the various user accounts. The user account 228a for user A and the user account 228n for user N are shown to include various information.


The information can include, without limitation, data associated with a display scroll order 230, user data, etc. As will be described in greater detail below, the display scroll order 230 includes information regarding a user's preferences, settings, and configurations which are settable by the user or set by default at the server 220 when accessing a respective user account. The storage 226 will include any number of user profiles, depending on the number of registered users having user accounts for their respective activity tracking devices. It should also be noted that a single user account can have various or multiple devices associated therewith, and the multiple devices can be individually customized, managed and accessed by a user. In one embodiment, the server 220 provides access to a user to view the user data 232 associated with activity tracking device.


The user data 232 viewable by the user includes the tracked motion data, which is processed to identify a plurality of metrics associated with the motion data. The user data 232 viewable by the user can include user heart beat and heart rate data 232A, which is processed to identify a plurality of metrics associated with the user's heart beat.


The metrics are shown in various graphical user interfaces of a website enabled by the server 220. The website can include various pages with graphical user interfaces for rendering and displaying the various metrics for view by the user associated with the user account. In one embodiment, the website can also include interfaces that allow for data entry and configuration by the user.


The configurations can include defining which metrics will be displayed on the activity tracking device 100. In addition, the configurations can include identification of which metrics will be a first metric to be displayed on the activity tracking device. The first metric to be displayed by the activity tracking device can be in response to a user input at the activity tracked device 100. As noted above, the user input can be by way of physical contact. The physical contact is qualified by the processor and/or logic of the activity tracking device 100 to determine if the physical contact should be treated as an input. The input can trigger or cause the display screen of the activity tracking device 100 to be turned on to display a specific metric, that is selected by the user as the first metric to display. In another embodiment, the first metric displayed in response to the input can be predefined by the system as a default.


The configuration provided by the user by way of the server 220 and the activity management application 224 can also be provided by way of the activity tracking application 202 of the computing device 200. For example, the activity tracking application 202 can include a plurality of screens that also display metrics associated with the captured motion data of the activity tracking device 100. The activity tracking application 202 can also allow for user input and configuration at various graphical user interface screens to set and define which input will produce display of the first metric. In other embodiments, in addition to identifying the first metric to be displayed in response to the input, which may be physical contact, the configuration can allow an ordering of which metrics will be displayed in a specific scroll order.


In another embodiment, the scroll order of the metrics is predefined. In some embodiments, the input provided by the user by way of the physical contact can be pre-assigned to a specific metric in the scroll order. For example, the scroll order can remain the same, while the input can allow the screen to jump to a specific entry in the scroll order. Jumping to a specific entry can be viewed as a shortcut to a specific entry that is desired to be seen first by the user upon providing physical contact or input to the device 100.



FIGS. 3A and 3B illustrate examples of activity tracking devices 300, 300′ having a heart rate measuring system in the form of a wearable wrist attachable device, in accordance with embodiments of the present invention. The form factor of the activity tracking devices 300, 300′ can be similar to the above activity tracking devices 100 and includes substantially similar components with the addition of the heart rate measuring system. The user's finger 302 is shown touching the activity tracking device at the heart rate monitor location 183. The light source is emitting light 320 into the user's skin 306. A portion 322 of the light 320 is reflected from the user's blood vessel 304.



FIGS. 3C and 3D illustrate another example of an activity tracking device 300″, in accordance with embodiments of the present invention. The form factor of the activity tracking device 300″ is shown as a clipable device that includes a spring loaded hinge 340, a screen 122, a button 126, and device components 102 integrated within the housing 190. The housing 190 can be a shape capable of receiving a user's finger 302 and pressing down on the finger from one or both sides of the finger. Inserting the user's finger 302 can provide the physical contact or tap needed to initiate certain functions of the tracking device 100 as will be described in more detail below.


The heart rate measuring system includes a light source 181 and a reflected light detector 182. The light source 181 and the reflected light detector 182 are located close together in the activity tracking devices 300, 300′, 300″. In one embodiment the light source 181 and the reflected light detector 182 can be immediately adjacent. The light source 181 and the reflected light detector 182 can be included in a single package and/or a single integrated circuit. The light source 181 and the reflected light detector 182 can be selected for any one suitable wavelength or suitable band of wavelengths of light ranging from between infrared, through a human visible spectrum to ultraviolet wavelengths. The heart rate monitor location 183 can include a cover that enables light of the light source to pass while blocking substantially all light in a human visible spectrum. The cover can be a smoked filter or other suitable filter color or shaded plastic or glass or shaded glass, transparent or translucent glass or plastic or ceramic or any other suitable material capable of allowing the desired wavelengths of light to pass through the cover. In one embodiment, the light source 181 uses an infrared (IR) light and the IR light produces a deadfront at the heart rate monitor location 183. Where a deadfront is defined as a continuous surface such that the cover is not easily discernable from the remaining surface of the housing. A deadfront cover is substantially hidden from the user however a light source 181 or the display screen 122 can emits sufficient light to pass through the cover. In another embodiment, the light source 181 can be a green light and the heart rate monitor location 183 can include a translucent green window.


In operation, the user places the skin 306 of a finger tip 302 or other body part over the light source 181. The light source 181 directs the light 310 into the skin 306. The light 310 passes through the skin 306 to a blood vessel 304 such as an artery, vein, or a capillary within the finger 302. A reflected portion 312 of the light 310 is reflected from the blood vessel 304 toward the reflected light detector 182. The light detector 182 outputs a signal corresponding to the reflected portion 312 of the light. The signal is coupled to a processor 106 for processing configured to identify heart beats of the user and produce an indication of a heart rate. The indication of the heart rate can be displayed on the display screen 122.


One embodiment may use a portion of the teachings of detecting heart beats by reflecting light from a blood vessel, as taught, in part by “Plug-and-Play, Single-Chip Photoplethysmography” by Deepak Chandrasekar, et al., pages 3243-3246, presented 34th Annual International Conference of the IEEE EMBS, San Diego, Calif. USA, 28 Aug.-1 Sep. 2012 which is incorporated by reference herein for all purposes.


Chandrasekar, et al, provides in pertinent part “a digital OPS can be used as a high-performance, reflectance-mode PPG sensor . . . LED emits light into the tissue, where it experiences diffuse reflection from the tissue and capillary bed. This establishes a baseline reflectance signal which is detected at the PD. When a pulse wave propagates through the capillary bed, the reflectance signal falls slightly (0.5-5%) due to light scattering. The change is detected by the PD and processed by embedded amplification and signal processing circuitry” (Page 3244, column 1, line 34 through column 2, line 7 and FIG. 2). Where an OPS is defined as an optical proximity sensor, a PPG sensor is defined as a photoplethysmographic sensor and a PD is defined as a photodiode.


It should be understood that the teachings described by are Chandrasekar, et al only examples and other examples can include different and additional processes and systems as described in more detail throughout this disclosure. Further, Chandrasekar, et al, fails to teach suitable filtering to provide accurate hear rate indications. Further still, Chandrasekar, et al, cannot discern motion caused false heart beat detections from actual hear beat detections.



FIG. 4A is a flowchart diagram of heart rate measuring system, in accordance with embodiments of the present invention. The heart rate measuring system detects the heart beats and interprets the detected heart beats to peaks. The detected peaks include timing information corresponding to the time interval between the detected peaks. The detected peaks also include information of the number of detected peaks within a selected sampling time period. The sampling time period can be determined by extent peaks such as the first detected peak and the last detected peak. The number of detected peaks within the sampling time period can be used to calculate an estimated heart rate in the form of beats per minute. The timing of the detected peaks can be used to evaluate the estimated heart rate to determine a best guess of the user's actual heart rate.



FIG. 4B is a graphical representation 420 of an example series of peaks corresponding to detected heart beats captured during a sampling time period STP, in accordance with embodiments of the present invention. The solid line peaks A, B, C, E, G and I the dashed peak F represent the raw data detected during the sampling time period STP.


The solid line peaks A, B, C, E, G and I correspond to actual detected heart beats. The dashed peak F corresponds to a phantom beat detection. The phantom peak F can be caused by movement of the user's finger or the activity tracking device 300. The phantom peak F can be caused by noise or some other cause.


The dotted peaks D and H correspond to approximate occurrences of heart beats that should have been detected, based on the timing of the actually detected peaks A, B, C, E, G and I. The dotted peaks D and H are not actually detected peaks and are shown in the graphical representation 420 for reference purposes as will be described in more detail below.


The raw heart beat data is next filtered to determine a best guess heart rate. FIG. 4C is a graphical representation 450 of the filtering process for the above example series of peaks, in accordance with embodiments of the present invention. The peaks A-I are transferred to FIG. 4C as a reference. Each of the horizontal lines of small circles represents a different estimate of peaks. Six actual peaks A, B, C, E, G and I correspond to six actual heart beats that were detected and one phantom peak F was detected giving a total of seven detected peaks.


Human heart beats are typically substantially evenly spaced, thus the seven detected peaks are separated by even time intervals across the sampling time period STP as shown in the estimated 7 peaks detected line. However, the filtering process evaluates the detection process to determine if the initial estimate of seven peaks is accurate as described in FIG. 5.



FIG. 5 is a flowchart diagram of the method operations 500 for detecting a heart rate, in accordance with embodiments of the present invention. In an operation 505, the user touches the heart rate monitor location 183 and/or presses the button 126 on the activity monitor to initiate the heart rate detection. The light source 181 and detector 182 are located inside the activity tracking device near the heart rate detector heart rate monitor location 183. The heart rate monitor location 183 can include a proximity sensor and/or the button 126 as described above.


In an operation 510, the light source 181 emits a light 320 into the user's skin and at least a portion of the light 322 is reflected off of a blood vessel 304 inside the user's skin. The detector 182 receives the reflected light 322 in an operation 515 and outputs the raw data of the detected beats within a sampling time period STP, as shown in FIG. 4B, to the processor 106 in an operation 520. The raw data of the detected beats is refined and evaluated to produce a more accurate heart rate.


As shown in FIG. 4C, the initial estimate of the seven detected heart beats are illustrated as seven, evenly spaced peaks. In an operation 525 at least one beat is added to the initial estimate of seven beats to produce a corresponding at least one added estimate. In the example above two peaks are added resulting in corresponding added estimates of an eight beat estimate and a nine beat estimate. It should be understood that only one or more than two beats could be added.


In an operation 530, at least one peak is subtracted from the initial estimate of seven beats to produce a corresponding at least one subtracted estimate. In the example above two beats are subtracted resulting in corresponding subtracted estimates of a six beat estimate and a five beat estimate. It should be understood that only one or more than two beats could be subtracted. As shown in the graphical representation 450, estimated lines of beats corresponding to five, six, eight and nine peaks are shown evenly distributed across the sampling time period STP.


In an operation 535, each of the five estimated lines of beats are compared to the actually detected peaks A, B, C, E, F, G and I to determine how well each line of beats scores as most closely matching the actually detected peaks A, B, C, E, F, G and I. By way of example, the five beats estimate line closely corresponds to only actually detected peak A and phantom peak F resulting in a corresponding score of 2. Further, the six beats estimate line closely corresponds to only actually detected peaks A, E and G resulting in a corresponding score of 3. Further, the initial estimate of seven beats estimate line closely corresponds to only actually detected peaks A, B and C and phantom peak F resulting in a corresponding score of 4. The eight beats estimate line closely corresponds to peaks A, B, C, E, G and I resulting in a corresponding score of 6. Finally, the nine beats estimate line closely corresponds to peaks A, B, C and I and phantom peak F resulting in a corresponding score of 5.


In an operation 540, the eight beats estimate provides the highest score of 6 and is therefore selected as a potential heart rate for further evaluation. However, only seven peaks were detected by the heart rate monitoring system and thus the eight beats potential heart rate could be an error.


In an operation 545, intervals between the seven detected beats are measured. Recall that human heart beats are substantially evenly spaced and there are three substantially evenly spaced peaks A, B, C in the seven detected beats. The time interval between the substantially evenly spaced peaks A, B, C is selected as a potential beat interval.


In an operation 550, the potential beat interval is compared to the remaining detected beat intervals between peaks C and E, between peaks E and F, between peaks F and G, and between peaks G and I to identify any timing intervals closely corresponding to whole number multiples of the potential beat interval so as to identify missed beats. In the above example, missed peak D is found between detected peaks C and E and missed peak H is found between detected peaks G and I. The timing intervals between peaks E and F and between peaks F and G are substantially less than the potential beat interval and thus peak F is identified as a phantom peak and the phantom peak F is eliminated in an operation 555.


In an operation 560, the estimated heart beats are scored similar to the scoring in operation 535 above. The combination of the actually detected peaks A, B, C, E, G and I and the missed peaks D and H can be evaluated for each of the estimated lines of beats. The eight beats potential heart rate would yield a resulting score of 8, which would be higher than the scores of 3, 2, 3, 4 corresponding to five, six, seven and nine beats estimate lines, respectively. The eight beats potential heart rate can therefore be determined as the best guess in an operation 565.


In an operation 570, the processor 106 calculates the heart rate. The sampling time period STP is extrapolated to 60 seconds and the corresponding heart rate in heart beats per minute is output to the display screen. By way of example, the sampling time period STP is 5.2 seconds in the above example. And 60 seconds divided by 5.2 seconds is equal to 11.54. And further, 11.54 multiplied by the seven time intervals between the evenly spaced estimated eight beats yields a calculated heart rate of 81 beats per minute. In an operation 575, the processor 106 outputs the calculated heart rate to the display screen 122.


It should be noted that the heart rate measuring system can continue to detect, refine and filter the detected heart beats as long as the user's skin is in sufficiently close proximity to the light source 181 and detector 182.



FIGS. 6A-D illustrate different presentations of the calculated heart rate on the display screen 122, in accordance with embodiments of the present invention. By way of example, in FIG. 6A the initially detected heart beats are displayed as peaks 602 and the heart rate 610 is displayed numerically “61 HRT”. As the user's finger remains on the activity tracking device the heart rate is more accurately detected and the displayed heart rate 610 is updated as shown in FIG. 6B. As shown in FIG. 6C the detected heart beats are displayed as lines 612 and as a sinewave 614 in FIG. 6D. It should be noted these examples are merely some examples of displaying the user's heart rate and others examples could include flashing an icon or a portion of or the entire the display 122 or vibrating the activity tracking device in time with the detected heart beats and combinations thereof.



FIG. 7A is a flowchart diagram of the method operations for adjusting the user pressure on the activity tracking device while detecting a heart rate, in accordance with embodiments of the present invention. FIGS. 7B and 7C show feedback signals on the activity tracing device display, in accordance with embodiments of the present invention. The button 126 can also include a pressure sensor that can be used to determine the user's pressure on the heart rate monitor location 183. In one embodiment, the pressure sensor can include at least one of a strain gauge, a push resistance built into the button 126, or a force sensitive film under, within or on a surface of the button and combinations thereof and any other suitable pressure sensor.


Alternatively, the excessive pressure or insufficient pressure can degrade the quality of the heart beat data in the reflected light 322 to a level that the heart rate cannot be suitably monitored. Similarly, the user's motion, such as running, can interfere with and produce artifacts and false readings. The processor 106 can use the motion sensor to identify and filter out the falsely detected heart beats from the detected heart beats. By way of example, the processor can detect when the user steps down and the vibration through the user's body and identify a detected heart beat coinciding with the step as a suspected false heart beat.


From operation 520 in FIG. 5 above, the processor may determine that the detected heart beats are too erratic to accurately measure in an operation 710. The detected heart beats are too erratic can be too erratic due to pressing the user's finger too lightly or too hard on the activity tracking device. In an operation 715, a feedback to the user is provided to increase pressure on the activity tracking device. As shown in FIG. 7B the increase pressure feedback can be in the form of one or more arrows pointing down or toward the user's finger. Alternatively, the activity device can flash a portion or all of the display 122 or display words such as “press down” or icons, or vibrate or other suitable tactile feedback, visual feedback, or audible feedback to the user.


The detected heart beats are constantly monitored for usability, in an operation 720. If increasing the pressure provides usable heart beat data then the method operations return to operation 525 in FIG. 5. If increasing the pressure does not provide usable heart beat data then the method operations continue in an operation 725 where the user is provided feedback to decrease the pressure on the activity tracking device. As shown in FIG. 7C the decrease pressure feedback can be in the form of one or more arrows pointing up or away from the user's finger. Alternatively, the activity device can flash a portion or all of the display 122 or display icons or words such as “lighten up” or vibrate or other suitable tactile feedback, visual feedback, or audible feedback to the user.


The detected heart beats are constantly monitored for usability, in an operation 730. If decreasing the pressure provides usable heart beat data then the method operations return to operation 525 in FIG. 5. If decreasing the pressure does not provide usable heart beat data then the method operations continue in operation 715 as described above.



FIG. 8 illustrates an example where various types of activities of users 800A-800I can be captured by activity tracking devices 100, in accordance with embodiments of the present invention. As shown, the various types of activities can generate different types of data that can be captured by the activity tracking device 100. The data, which can be represented as motion data (or processed motion data) can be transferred 820 to a network 176 for processing and saving by a server, as described above. In one embodiment, the activity tracking device 100 can communicate to a device using a wireless connection, and the device is capable of communicating and synchronizing the captured data with an application running on the server. In one embodiment, an application running on a local device, such as a smart phone or tablet or smart watch can capture or receive data from the activity tracking device 100 and represent the tract motion data in a number of metrics.


In one embodiment, the device collects one or more types of physiological and/or environmental data from embedded sensors and/or external devices and communicates or relays such metric information to other devices, including devices capable of serving as Internet-accessible data sources, thus permitting the collected data to be viewed, for example, using a web browser or network-based application. For example, while the user is wearing an activity tracking device, the device may calculate and store the user's step count using one or more sensors. The device then transmits data representative of the user's step count to an account on a web service, computer, mobile phone, or health station where the data may be stored, processed, and visualized by the user. Indeed, the device may measure or calculate a plurality of other physiological metrics in addition to, or in place of, the user's step count.


Some physiological metrics include, but are not limited to, energy expenditure (for example, calorie burn), floors climbed and/or descended, heart rate, heart rate variability, heart rate recovery, location and/or heading (for example, through GPS), elevation, ambulatory speed and/or distance traveled, swimming lap count, bicycle distance and/or speed, blood pressure, blood glucose, skin conduction, skin and/or body temperature, electromyography, electroencephalography, weight, body fat, caloric intake, nutritional intake from food, medication intake, sleep periods (i.e., clock time), sleep phases, sleep quality and/or duration, pH levels, hydration levels, and respiration rate. The device may also measure or calculate metrics related to the environment around the user such as barometric pressure, weather conditions (for example, temperature, humidity, pollen count, air quality, rain/snow conditions, wind speed), light exposure (for example, ambient light, UV light exposure, time and/or duration spent in darkness), noise exposure, radiation exposure, and magnetic field.


Still further, other metrics can include, without limitation, calories burned by a user, weight gained by a user, weight lost by a user, stairs ascended, e.g., climbed, etc., by a user, stairs descended by a user, variation in the user's altitude, steps taken by a user during walking or running, a number of rotations of a bicycle pedal rotated by a user, sedentary activity data, driving a vehicle, a number of golf swings taken by a user, a number of forehands of a sport played by a user, a number of backhands of a sport played by a user, or a combination thereof. In some embodiments, sedentary activity data is referred to herein as inactive activity data or as passive activity data. In some embodiments, when a user is not sedentary and is not sleeping, the user is active. In some embodiments, a user may stand on a monitoring device that determines a physiological parameter of the user. For example, a user stands on a scale that measures a weight, a body fat percentage, a biomass index, or a combination thereof, of the user.


Furthermore, the device or the system collating the data streams may calculate metrics derived from this data. For example, the device or system may calculate the user's stress and/or relaxation levels through a combination of heart rate variability, skin conduction, noise pollution, and sleep quality. In another example, the device or system may determine the efficacy of a medical intervention (for example, medication) through the combination of medication intake, sleep and/or activity data. In yet another example, the device or system may determine the efficacy of an allergy medication through the combination of pollen data, medication intake, sleep and/or activity data. These examples are provided for illustration only and are not intended to be limiting or exhaustive.


This information can be associated to the users account, which can be managed by an activity management application on the server. The activity management application can provide access to the users account and data saved thereon. The activity manager application running on the server can be in the form of a web application. The web application can provide access to a number of websites screens and pages that illustrate information regarding the metrics in various formats. This information can be viewed by the user, and synchronized with a computing device of the user, such as a smart phone.


In one embodiment, the data captured by the activity tracking device 100 is received by the computing device, and the data is synchronized with the activity measured application on the server. In this example, data viewable on the computing device (e.g. smart phone) using an activity tracking application (app) can be synchronized with the data present on the server, and associated with the user's account. In this way, information entered into the activity tracking application on the computing device can be synchronized with application illustrated in the various screens of the activity management application provided by the server on the website.


The user can therefore access the data associated with the user account using any device having access to the Internet. Data received by the network 176 can then be synchronized with the user's various devices, and analytics on the server can provide data analysis to provide recommendations for additional activity, and or improvements in physical health. The process therefore continues where data is captured, analyzed, synchronized, and recommendations are produced. In some embodiments, the captured data can be itemized and partitioned based on the type of activity being performed, and such information can be provided to the user on the website via graphical user interfaces, or by way of the application executed on the users smart phone (by way of graphical user interfaces).


In an embodiment, the sensor or sensors of a device 100 can determine or capture data to determine an amount of movement of the monitoring device over a period of time. The sensors can include, for example, an accelerometer, a magnetometer, a gyroscope, or combinations thereof. Broadly speaking, these sensors are inertial sensors, which capture some movement data, in response to the device 100 being moved. The amount of movement (e.g., motion sensed) may occur when the user is performing an activity of climbing stairs over the time period, walking, running, etc. The monitoring device may be worn on a wrist, carried by a user, worn on clothing (using a clip, or placed in a pocket), attached to a leg or foot, attached to the user's chest, waist, or integrated in an article of clothing such as a shirt, hat, pants, blouse, glasses, and the like. These examples are not limiting to all the possible ways the sensors of the device can be associated with a user or thing being monitored.


In other embodiments, a biological sensor can determine any number of physiological characteristics of a user. As another example, the biological sensor may determine heart rate, a hydration level, body fat, bone density, fingerprint data, sweat rate, and/or a bioimpedance of the user. Examples of the biological sensors include, without limitation, a biometric sensor, a physiological parameter sensor, a pedometer, or a combination thereof.


In some embodiments, data associated with the user's activity can be monitored by the applications on the server and the users device, and activity associated with the user's friends, acquaintances, or social network peers can also be shared, based on the user's authorization. This provides for the ability for friends to compete regarding their fitness, achieve goals, receive badges for achieving goals, get reminders for achieving such goals, rewards or discounts for achieving certain goals, etc.


In some embodiments the heart beats are detected by processing the light received in the light detector, within the sampling time period. A first time interval between a first beat of the detected heart beats and a second heart beat is measured. The sample time interval can be divided by the first time interval to determine a first estimate of heart beats detected within the sampling time period and the first estimate of heart beats within the sampling time period can be extrapolated to a first estimated heart beats per minute. The first estimate heart beats per minute can be output to the display screen. At least one beat can be added to the first estimate of heart beats to produce a second estimate of heart beats that can be scored with the first estimate of heart beats. A highest scoring of the first estimate of heart beats and the second estimate of heart beats can be selected and output to the display screen.


In another embodiment, at least one beat can be subtracted from the first estimate of beats to produce a third estimate of heart beats. The first estimate of heart beats and the third estimate of heart beats can be scored. A highest scoring of the first estimate of heart beats and the third estimate of heart beats can be selected and output to the display screen.


In another embodiment, identifying the heart beats of the user and producing an indication of a heart rate can include identifying and filtering a falsely detected heart beat coinciding with motion detected by the motion sensor. By way of example, a user's motion may be erroneously identified as a heart beat. The processor can compare detected motion (i.e., motion data, instantaneous shocks, etc.) to the detected heart beats and identify heart beats that coincide with motion data. Further, as the motion data and heart beat data are compiled over time, detected motions that often produce corresponding erroneously detected heart beats can be identified and filtered from the detected heart beats.


In another embodiment, activity tracking device includes a housing including a motion sensor and a processor. The processor is configured for processing motion data produced by the motion sensor. A display screen is integrated with the housing to display metrics that quantify the motion data produced by the motion sensor. A light source is integrated within the housing to enable light to be directed out of the housing at a heart rate monitor location on the housing. A light detector is also integrated within the housing. The light detector is configured to capture an amount of the light that is reflected back to the light detector, at least a first portion of the light reflected back to the light detector is reflected from a blood vessel disposed under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor can be in communication with the light detector to enable processing of the reflected light to identify heart beats of the user and produce an indication of a heart rate. The indication of the heart rate being displayable on the display screen as an option, in addition to the metrics that quantify the motion data. The heart rate can be calculated based on an algorithm that calculates a first estimate of heart beats per minute corresponding to detected heart beats in the light received in the light detector within a sampling time period. A refined estimate of heart beats per minute can be calculated by adding at least one beat to the first estimate of heart beats to produce a second estimate of heart beats and subtracting at least one beat from the first estimate of beats to produce a third estimate of heart beats. The first estimate of heart beats, the second estimate of heart beats and the third estimate of heart beats are scored and a highest scoring estimate of heart beats is selected and output to the display screen.


As noted, an activity tracking device 100 can communicate with a computing device (e.g., a smartphone, a tablet computer, a desktop computer, or computer device having wireless communication access and/or access to the Internet). The computing device, in turn, can communicate over a network, such as the Internet or an Intranet to provide data synchronization. The network may be a wide area network, a local area network, or a combination thereof. The network may be coupled to one or more servers, one or more virtual machines, or a combination thereof. A server, a virtual machine, a controller of a monitoring device, or a controller of a computing device is sometimes referred to herein as a computing resource. Examples of a controller include a processor and a memory device.


In one embodiment, the processor may be a general purpose processor. In another embodiment, the processor can be a customized processor configured to run specific algorithms or operations. Such processors can include digital signal processors (DSPs), which are designed to execute or interact with specific chips, signals, wires, and perform certain algorithms, processes, state diagrams, feedback, detection, execution, or the like. In some embodiments, a processor can include or be interfaced with an application specific integrated circuit (ASIC), a programmable logic device (PLD), a central processing unit (CPU), or a combination thereof, etc.


In some embodiments, one or more chips, modules, devices, or logic can be defined to execute instructions or logic, which collectively can be viewed or characterized to be a processor. Therefore, it should be understood that a processor does not necessarily have to be one single chip or module, but can be defined from a collection of electronic or connecting components, logic, firmware, code, and combinations thereof.


Examples of a memory device include a random access memory (RAM) and a read-only memory (ROM). A memory device may be a Flash memory, a redundant array of disks (RAID), a hard disk, or a combination thereof.


Embodiments described in the present disclosure may be practiced with various computer system configurations including hand-held devices, microprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers and the like. Several embodiments described in the present disclosure can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a wire-based or wireless network.


With the above embodiments in mind, it should be understood that a number of embodiments described in the present disclosure can employ various computer-implemented operations involving data stored in computer systems. These operations are those requiring physical manipulation of physical quantities. Any of the operations described herein that form part of various embodiments described in the present disclosure are useful machine operations. Several embodiments described in the present disclosure also relate to a device or an apparatus for performing these operations. The apparatus can be specially constructed for a purpose, or the apparatus can be a computer selectively activated or configured by a computer program stored in the computer. In particular, various machines can be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.


Various embodiments described in the present disclosure can also be embodied as computer-readable code on a non-transitory computer-readable medium. The computer-readable medium is any data storage device that can store data, which can thereafter be read by a computer system. Examples of the computer-readable medium include hard drives, network attached storage (NAS), ROM, RAM, compact disc-ROMs (CD-ROMs), CD-recordables (CD-Rs), CD-rewritables (RWs), magnetic tapes and other optical and non-optical data storage devices. The computer-readable medium can include computer-readable tangible medium distributed over a network-coupled computer system so that the computer-readable code is stored and executed in a distributed fashion.


Although the method operations were described in a specific order, it should be understood that other housekeeping operations may be performed in between operations, or operations may be performed in an order other than that shown, or operations may be adjusted so that they occur at slightly different times, or may be distributed in a system which allows the occurrence of the processing operations at various intervals associated with the processing.


Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications can be practiced within the scope of the appended claims. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the various embodiments described in the present disclosure are not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.

Claims
  • 1. An activity tracking device, comprising: a light source configured to emit light;a light detector configured to detect light emitted from the light source and reflected from a blood vessel of a user;a display screen;at least one processor; anda memory storing computer-executable instructions for controlling the at least one processor to: determine an amount of light received by the light detector based on a signal output from the light detector, at least a portion of the determined amount of light being reflected from the blood vessel,determine a baseline reflectance signal of the light reflected from the blood vessel between heart beats of the user based on the amount of light received by the light detector,determine a second reflectance signal corresponding to at least one of the heart beats of the user based on the amount of light received by the light detector, the second reflectance signal being less than the first reflectance signal,identify the heart beats of the user within a sampling time period based on the first and second reflectance signals,measure a first time interval between a first heart beat of the identified heart beats and a second heart beat of the identified heart beats;divide the sampling time period by the first time interval to determine a first estimate of the heart beats within the sampling time period;add at least one heart beat to the first estimate of the heart beats to produce a second estimate of heart beats within the sampling time period,subtract at least one heart beat from the first estimate of the heart beats to produce a third estimate of heart beats within the sampling time period,score the first estimate of the heart beats, the second estimate of the heart beats, and the third estimate of the heart beats,select a highest-scoring one of the first estimate of the heart beats, the second estimate of the heart beats, and the third estimate of the heart beats as a selected estimate of the heart beats,estimating a first heart beats per minute at least in part, on the selected estimate of the heart beats; andcause the estimated first heart beats per minute to be displayed via the display screen.
  • 2. The activity tracking device of claim 1, wherein the memory further stores computer-executable instructions for controlling the at least one processor to identify the heart beats based on differentiating between the baseline reflectance signal and the second reflectance signal.
  • 3. The activity tracking device of claim 1, further comprising at least one environmental sensor configured to detect motion of the activity tracking device, wherein the memory further stores computer-executable instructions for controlling the at least one processor to: identify at least one falsely detected heart beat coinciding with motion detected by the at least one environmental sensor, andfilter the at least one identified falsely detected heart beat.
  • 4. The activity tracking device of claim 1, further comprising a pressure sensor adjacent to or collocated with one or both of the light detector and the light source, wherein the memory further stores computer-executable instructions for controlling the at least one processor to: detect a pressure via the pressure sensor during the identification of the heart beats, anddisplay a feedback signal regarding the detected pressure, the feedback signal being indicative of a need for the application of more or less pressure during heart rate detection.
  • 5. A method operable by a processor of an activity tracking device, comprising: determining an amount of light received by a light detector of the activity tracking device based on a signal output from the light detector, at least a first portion of the determined amount of light being reflected from a blood vessel disposed under a skin of a user, the activity tracking device comprising a light source configured to emit light onto the blood vessel to be reflected by the blood vessel;determining a baseline reflectance signal of the light reflected from the blood vessel between heart beats of the user based on the amount of light received by the light detector;determining a second reflectance signal corresponding to at least one of the heart beats of the user based on the amount of light received by the light detector, the second reflectance signal being less than the baseline reflectance signal;identifying the heart beats of the user within a sampling time period based on the first and second reflectance signals;measuring a first time interval between a first beat of the identified heart beats and a second beat of the identified heart beats;dividing the sampling time period by the first time interval to determine a first estimate of the heart beats within the sampling time period;adding at least one heart beat to the first estimate of the heart beats to produce a second estimate of the heart beats within the sampling time period;subtracting at least one heart beat from the first estimate of beats to produce a third estimate of the heart beats within the sampling time period;scoring the first estimate of the heart beats, the second estimate of the heart beats and the third estimate of the heart beats;selecting a highest-scoring one of the first estimate of the heart beats, the second estimate of the heart beats, and the third estimate of the heart beats as the selected estimated the heart beats;estimating a first heart beats per minute based, at least in part, on the selected estimate of the heart beats; andcausing the estimated first heart beats per minute to be displayed via the display screen.
  • 6. The method of claim 5, further comprising identifying the heart beats based on differentiating between the first reflectance signal and the second reflectance signal.
  • 7. The method of claim 5, further comprising: identifying at least one falsely detected heart beat coinciding with motion detected by at least one environmental sensor configured to detect motion of the activity tracking device; andfiltering the at least one identified falsely detected heart beat.
  • 8. The method of claim 5, further comprising: detecting a pressure via a pressure sensor of the activity tracking device during the identification of the heart beats, the pressure sensor located adjacent to or collocated with one or both of the light detector and the light source; anddisplaying a feedback signal regarding the detected pressure, the feedback signal being indicative of a need for the application of more or less pressure during heart rate detection.
CROSS REFERENCE TO RELATED APPLICATIONS

This application is a continuation of and claims priority from U.S. patent application Ser. No. 14/639,409, filed on Mar. 5, 2015 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety and is a continuation of U.S. patent application Ser. No. 14/302,360 (granted as U.S. Pat. No. 9,039,614), filed on Jun. 11, 2014 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety. U.S. patent application Ser. No. 14/302,360 is a continuation of and claims priority from U.S. patent application Ser. No. 14/156,381 (granted as U.S. Pat. No. 8,827,906), filed on Jan. 15, 2014 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety. This application also claims priority, through U.S. Pat. No. 8,827,906, from U.S. Provisional Patent Application No. 61/924,547 filed on Jan. 7, 2014 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety. This application also claims priority, through U.S. Pat. No. 8,827,906, from U.S. Provisional Patent Application No. 61/752,826 filed on Jan. 15, 2013 and entitled “Portable Monitoring Devices and Methods of Operating Same,” which is incorporated herein by reference in its entirety. This application also claims priority through U.S. Pat. No. 8,827,906, from U.S. Provisional Patent Application No. 61/830,600 filed on Jun. 3, 2013 and entitled “Portable Monitoring Devices and Methods of Operating Same,” which is incorporated herein by reference in its entirety.

US Referenced Citations (648)
Number Name Date Kind
2717736 Schlesinger Sep 1955 A
2827309 Fred Mar 1958 A
2883255 Anderson Apr 1959 A
3163856 Kirby Dec 1964 A
3250270 Walter May 1966 A
3522383 Chang Jul 1970 A
3608545 Novack et al. Sep 1971 A
3918658 Beller Nov 1975 A
4192000 Lipsey Mar 1980 A
4244020 Ratcliff Jan 1981 A
4258719 Lewyn Mar 1981 A
4281663 Pringle Aug 1981 A
4284849 Anderson et al. Aug 1981 A
4312358 Barney Jan 1982 A
4367752 Jimenez et al. Jan 1983 A
4390922 Pelliccia Jun 1983 A
4407295 Steuer et al. Oct 1983 A
4425921 Fujisaki et al. Jan 1984 A
4575804 Ratcliff Mar 1986 A
4578769 Frederick Mar 1986 A
4617525 Lloyd Oct 1986 A
4771792 Seale Sep 1988 A
4781195 Martin Nov 1988 A
4846183 Martin Jul 1989 A
4887249 Thinesen Dec 1989 A
4930518 Hrushesky et al. Jun 1990 A
4960126 Conlon et al. Oct 1990 A
4977509 Pitchford et al. Dec 1990 A
5036856 Thornton Aug 1991 A
5058427 Brandt Oct 1991 A
5101831 Koyama et al. Apr 1992 A
5224059 Nitta et al. Jun 1993 A
5295085 Hoffacker Mar 1994 A
5301154 Suga Apr 1994 A
5314389 Dotan May 1994 A
5318597 Hauck et al. Jun 1994 A
5323650 Fullen et al. Jun 1994 A
5365930 Takashima et al. Nov 1994 A
5446705 Haas et al. Aug 1995 A
5456648 Edinburg et al. Oct 1995 A
5490523 Isaacson et al. Feb 1996 A
5513649 Gevins et al. May 1996 A
5553296 Forrest et al. Sep 1996 A
5583776 Levi et al. Dec 1996 A
5645509 Brewer et al. Jul 1997 A
5671162 Werbin Sep 1997 A
5704350 Williams, III Jan 1998 A
5724265 Hutchings Mar 1998 A
5734625 Kondo Mar 1998 A
5738104 Lo et al. Apr 1998 A
5817008 Ratert et al. Oct 1998 A
5830137 Scharf Nov 1998 A
5890128 Diaz et al. Mar 1999 A
5891042 Sham et al. Apr 1999 A
5894454 Kondo Apr 1999 A
5899963 Hutchings May 1999 A
5941828 Archibald et al. Aug 1999 A
5947868 Dugan Sep 1999 A
5954644 Dettling et al. Sep 1999 A
5955667 Fyfe Sep 1999 A
5976083 Richardson et al. Nov 1999 A
6018705 Gaudet et al. Jan 2000 A
6076015 Hartley et al. Jun 2000 A
6077193 Buhler et al. Jun 2000 A
6078874 Piety et al. Jun 2000 A
6085248 Sambamurthy et al. Jul 2000 A
6099478 Aoshima et al. Aug 2000 A
6129686 Friedman Oct 2000 A
6131076 Stephan et al. Oct 2000 A
6145389 Ebeling et al. Nov 2000 A
6183425 Whalen et al. Feb 2001 B1
6213872 Harada et al. Apr 2001 B1
6241684 Amano et al. Jun 2001 B1
6287262 Amano et al. Sep 2001 B1
6289230 Chaiken et al. Sep 2001 B1
6301964 Fyfe et al. Oct 2001 B1
6302789 Harada et al. Oct 2001 B2
6305221 Hutchings Oct 2001 B1
6307576 Rosenfeld Oct 2001 B1
6309360 Mault Oct 2001 B1
6360113 Dettling Mar 2002 B1
6402690 Rhee et al. Jun 2002 B1
6418394 Puolakanaho et al. Jul 2002 B1
6469639 Tanenhaus et al. Oct 2002 B2
6478736 Mault Nov 2002 B1
6513381 Fyfe et al. Feb 2003 B2
6513532 Mault et al. Feb 2003 B2
6527711 Stivoric et al. Mar 2003 B1
6529827 Beason et al. Mar 2003 B1
6558335 Thede May 2003 B1
6561951 Cannon et al. May 2003 B2
6571200 Mault May 2003 B1
6583369 Montagnino et al. Jun 2003 B2
6585622 Shum et al. Jul 2003 B1
6607493 Song Aug 2003 B2
6620078 Pfeffer Sep 2003 B2
6678629 Tsuji Jan 2004 B2
6699188 Wessel Mar 2004 B2
6720860 Narayanaswami Apr 2004 B1
6731967 Turcott May 2004 B1
6761064 Tsuji Jul 2004 B2
6790178 Mault et al. Sep 2004 B1
6808473 Hisano et al. Oct 2004 B2
6811516 Dugan Nov 2004 B1
6813582 Levi et al. Nov 2004 B2
6813931 Yadav et al. Nov 2004 B2
6856938 Kurtz Feb 2005 B2
6862575 Anttila et al. Mar 2005 B1
6882955 Ohlenbusch et al. Apr 2005 B1
6959259 Vock et al. Oct 2005 B2
6997882 Parker et al. Feb 2006 B1
7020508 Stivoric et al. Mar 2006 B2
7041032 Calvano May 2006 B1
7062225 White Jun 2006 B2
7099237 Lall Aug 2006 B2
7133690 Ranta-Aho et al. Nov 2006 B2
7153262 Stivoric et al. Dec 2006 B2
7162368 Levi et al. Jan 2007 B2
7171331 Vock et al. Jan 2007 B2
7200517 Darley et al. Apr 2007 B2
7246033 Kudo Jul 2007 B1
7252639 Kimura et al. Aug 2007 B2
7261690 Teller et al. Aug 2007 B2
7272982 Neuhauser et al. Sep 2007 B2
7285090 Stivoric et al. Oct 2007 B2
7334472 Seo et al. Feb 2008 B2
7373820 James May 2008 B1
7443292 Jensen et al. Oct 2008 B2
7457724 Vock et al. Nov 2008 B2
7467060 Kulach et al. Dec 2008 B2
7502643 Farringdon et al. Mar 2009 B2
7505865 Ohkubo et al. Mar 2009 B2
7539532 Tran May 2009 B2
7558622 Tran Jul 2009 B2
7559877 Parks et al. Jul 2009 B2
7579946 Case, Jr. Aug 2009 B2
7608050 Shugg Oct 2009 B2
7653508 Kahn et al. Jan 2010 B1
7690556 Kahn et al. Apr 2010 B1
7713173 Shin et al. May 2010 B2
7720306 Gardiner et al. May 2010 B2
7762952 Lee et al. Jul 2010 B2
7771320 Riley et al. Aug 2010 B2
7774156 Niva et al. Aug 2010 B2
7789802 Lee et al. Sep 2010 B2
7881902 Kahn et al. Feb 2011 B1
7909768 Turcott Mar 2011 B1
7927253 Vincent et al. Apr 2011 B2
7942824 Kayyali et al. May 2011 B1
7953549 Graham May 2011 B2
7983876 Vock et al. Jul 2011 B2
7993276 Nazarian et al. Aug 2011 B2
8005922 Boudreau Aug 2011 B2
8028443 Case, Jr. Oct 2011 B2
8040758 Dickinson Oct 2011 B1
8055469 Kulach et al. Nov 2011 B2
8073707 Teller et al. Dec 2011 B2
8099318 Moukas et al. Jan 2012 B2
8109858 Redmann Feb 2012 B2
8132037 Fehr et al. Mar 2012 B2
8140143 Picard et al. Mar 2012 B2
8152745 Smith et al. Apr 2012 B2
8157731 Teller et al. Apr 2012 B2
8172761 Rulkov et al. May 2012 B1
8177260 Tropper et al. May 2012 B2
8180591 Yuen et al. May 2012 B2
8180592 Yuen et al. May 2012 B2
8199126 Taubman Jun 2012 B1
8211503 Tsao et al. Jul 2012 B2
8270297 Akasaka et al. Sep 2012 B2
8311769 Yuen et al. Nov 2012 B2
8311770 Yuen et al. Nov 2012 B2
8321004 Moon et al. Nov 2012 B2
8346328 Mannheimer et al. Jan 2013 B2
8364250 Moon et al. Jan 2013 B2
8386008 Yuen et al. Feb 2013 B2
8386042 Yudovsky et al. Feb 2013 B2
8398546 Pacione et al. Mar 2013 B2
8437824 Moon et al. May 2013 B2
8437980 Yuen et al. May 2013 B2
8444578 Bourget et al. May 2013 B2
8446275 Utter, II May 2013 B2
8462591 Marhaben Jun 2013 B1
8463576 Yuen et al. Jun 2013 B2
8463577 Yuen et al. Jun 2013 B2
8475367 Yuen et al. Jul 2013 B1
8487771 Hsieh et al. Jul 2013 B2
8527038 Moon et al. Sep 2013 B2
8533269 Brown Sep 2013 B2
8533620 Hoffman et al. Sep 2013 B2
8543185 Yuen et al. Sep 2013 B2
8543351 Yuen et al. Sep 2013 B2
8548770 Yuen et al. Oct 2013 B2
8554297 Moon et al. Oct 2013 B2
8562489 Burton et al. Oct 2013 B2
8579827 Rulkov et al. Nov 2013 B1
8583402 Yuen et al. Nov 2013 B2
8597093 Engelberg et al. Dec 2013 B2
8634796 Johnson Jan 2014 B2
8641612 Teller et al. Feb 2014 B2
8670953 Yuen et al. Mar 2014 B2
8684900 Tran Apr 2014 B2
8690578 Nusbaum et al. Apr 2014 B1
8738118 Moon et al. May 2014 B2
8738321 Yuen et al. May 2014 B2
8738323 Yuen et al. May 2014 B2
8742325 Droz et al. Jun 2014 B1
8744803 Park et al. Jun 2014 B2
8762101 Yuen et al. Jun 2014 B2
8764651 Tran Jul 2014 B2
8792981 Yudovsky et al. Jul 2014 B2
8827906 Yuen et al. Sep 2014 B2
8847988 Geisner et al. Sep 2014 B2
8868377 Yuen et al. Oct 2014 B2
8909543 Tropper et al. Dec 2014 B2
8920332 Hong et al. Dec 2014 B2
8936552 Kateraas et al. Jan 2015 B2
8938279 Heaton, II et al. Jan 2015 B1
8945017 Venkatraman et al. Feb 2015 B2
8948832 Hong et al. Feb 2015 B2
8949070 Kahn et al. Feb 2015 B1
8954135 Yuen et al. Feb 2015 B2
8954290 Yuen et al. Feb 2015 B2
8956303 Hong et al. Feb 2015 B2
8961413 Teller et al. Feb 2015 B2
8968195 Tran Mar 2015 B2
8998815 Venkatraman et al. Apr 2015 B2
9005129 Venkatraman et al. Apr 2015 B2
9014790 Richards et al. Apr 2015 B2
9031812 Roberts et al. May 2015 B2
9039614 Yuen et al. May 2015 B2
9042971 Brumback et al. May 2015 B2
9044149 Richards et al. Jun 2015 B2
9044150 Brumback et al. Jun 2015 B2
9047648 Lekutai et al. Jun 2015 B1
9049998 Brumback et al. Jun 2015 B2
9089760 Tropper et al. Jul 2015 B2
9113794 Hong et al. Aug 2015 B2
9113795 Hong et al. Aug 2015 B2
9226663 Fei Jan 2016 B2
9237855 Hong et al. Jan 2016 B2
9241646 Addison et al. Jan 2016 B2
9282902 Richards et al. Mar 2016 B2
9307917 Hong et al. Apr 2016 B2
9314166 Brady et al. Apr 2016 B1
9314197 Eisen et al. Apr 2016 B2
9392946 Sarantos et al. Jul 2016 B1
9402552 Richards et al. Aug 2016 B2
9456787 Venkatraman et al. Oct 2016 B2
9596999 Moon Mar 2017 B2
9662053 Richards et al. May 2017 B2
9775548 Sarantos et al. Oct 2017 B2
10178973 Venkatraman et al. Jan 2019 B2
10216893 Hong et al. Feb 2019 B2
10216894 Hong et al. Feb 2019 B2
10381109 Hong et al. Aug 2019 B2
10433739 Weekly et al. Oct 2019 B2
10512407 Richards et al. Dec 2019 B2
10568525 Wu et al. Feb 2020 B1
20010044588 Mault Nov 2001 A1
20010049470 Mault et al. Dec 2001 A1
20010055242 Deshmuhk et al. Dec 2001 A1
20020013717 Ando et al. Jan 2002 A1
20020019585 Dickinson Feb 2002 A1
20020077219 Cohen et al. Jun 2002 A1
20020077536 Diab et al. Jun 2002 A1
20020082144 Pfeffer Jun 2002 A1
20020087264 Hills et al. Jul 2002 A1
20020091329 Heikkila et al. Jul 2002 A1
20020109600 Mault et al. Aug 2002 A1
20020139936 Dumas Oct 2002 A1
20020178060 Sheehan Nov 2002 A1
20020198776 Nara et al. Dec 2002 A1
20030018523 Rappaport et al. Jan 2003 A1
20030050537 Wessel Mar 2003 A1
20030065561 Brown et al. Apr 2003 A1
20030107487 Korman et al. Jun 2003 A1
20030128867 Bennett Jul 2003 A1
20030131059 Brown et al. Jul 2003 A1
20030163710 Ortiz et al. Aug 2003 A1
20030171189 Kaufman Sep 2003 A1
20030229276 Sarussi et al. Dec 2003 A1
20040054497 Kurtz Mar 2004 A1
20040061324 Howard Apr 2004 A1
20040117963 Schneider Jun 2004 A1
20040152957 Stivoric et al. Aug 2004 A1
20040171969 Socci et al. Sep 2004 A1
20040190085 Silverbrook et al. Sep 2004 A1
20040236227 Gueissaz Nov 2004 A1
20040239497 Schwartzman et al. Dec 2004 A1
20040249299 Cobb Dec 2004 A1
20040257557 Block Dec 2004 A1
20050020927 Blondeau et al. Jan 2005 A1
20050037844 Shum et al. Feb 2005 A1
20050038679 Short Feb 2005 A1
20050054938 Wehman et al. Mar 2005 A1
20050054940 Almen Mar 2005 A1
20050102172 Sirmans, Jr. May 2005 A1
20050107723 Wehman et al. May 2005 A1
20050163056 Ranta-Aho et al. Jul 2005 A1
20050171410 Hjelt et al. Aug 2005 A1
20050186965 Pagonis et al. Aug 2005 A1
20050187481 Hatib et al. Aug 2005 A1
20050195830 Chitrapu et al. Sep 2005 A1
20050228244 Banet Oct 2005 A1
20050228692 Hodgdon Oct 2005 A1
20050234742 Hodgdon Oct 2005 A1
20050245793 Hilton et al. Nov 2005 A1
20050248718 Howell et al. Nov 2005 A1
20050253047 Maegawa et al. Nov 2005 A1
20050272564 Pyles et al. Dec 2005 A1
20060004265 Pulkkinen et al. Jan 2006 A1
20060020174 Matsumura Jan 2006 A1
20060020177 Seo et al. Jan 2006 A1
20060025282 Redmann Feb 2006 A1
20060039348 Racz et al. Feb 2006 A1
20060047208 Yoon Mar 2006 A1
20060047447 Brady et al. Mar 2006 A1
20060052727 Palestrant Mar 2006 A1
20060064276 Ren et al. Mar 2006 A1
20060069619 Walker et al. Mar 2006 A1
20060089542 Sands Apr 2006 A1
20060111944 Sirmans, Jr. May 2006 A1
20060129436 Short Jun 2006 A1
20060143645 Vock et al. Jun 2006 A1
20060166718 Seshadri et al. Jul 2006 A1
20060189872 Arnold Aug 2006 A1
20060195020 Martin et al. Aug 2006 A1
20060217231 Parks et al. Sep 2006 A1
20060247952 Muraca Nov 2006 A1
20060277474 Robarts et al. Dec 2006 A1
20060282021 DeVaul et al. Dec 2006 A1
20060287883 Turgiss et al. Dec 2006 A1
20060288117 Raveendran Dec 2006 A1
20070011028 Sweeney Jan 2007 A1
20070049384 Kinq et al. Mar 2007 A1
20070050715 Behar Mar 2007 A1
20070051369 Choi et al. Mar 2007 A1
20070061593 Celikkan et al. Mar 2007 A1
20070071643 Hall et al. Mar 2007 A1
20070072156 Kaufman et al. Mar 2007 A1
20070083095 Rippo et al. Apr 2007 A1
20070083602 Heggenhougen et al. Apr 2007 A1
20070123391 Shin et al. May 2007 A1
20070135264 Rosenberg Jun 2007 A1
20070136093 Rankin et al. Jun 2007 A1
20070146116 Kimbrell Jun 2007 A1
20070155277 Amitai et al. Jul 2007 A1
20070159926 Prstojevich et al. Jul 2007 A1
20070179356 Wessel Aug 2007 A1
20070194066 Ishihara et al. Aug 2007 A1
20070197920 Adams Aug 2007 A1
20070208544 Kulach et al. Sep 2007 A1
20070213020 Novac Sep 2007 A1
20070219059 Schwartz et al. Sep 2007 A1
20070265533 Tran Nov 2007 A1
20070276271 Chan Nov 2007 A1
20070288265 Quinian et al. Dec 2007 A1
20080001735 Tran Jan 2008 A1
20080014947 Carnall Jan 2008 A1
20080022089 Leedom Jan 2008 A1
20080032864 Hakki Feb 2008 A1
20080039729 Cho et al. Feb 2008 A1
20080054072 Katragadda Mar 2008 A1
20080084823 Akasaka et al. Apr 2008 A1
20080093838 Tropper et al. Apr 2008 A1
20080097221 Florian Apr 2008 A1
20080097550 Dicks et al. Apr 2008 A1
20080114829 Button et al. May 2008 A1
20080125288 Case May 2008 A1
20080129457 Ritter et al. Jun 2008 A1
20080134102 Movold et al. Jun 2008 A1
20080140163 Keacher et al. Jun 2008 A1
20080140338 No et al. Jun 2008 A1
20080146892 LeBoeuf et al. Jun 2008 A1
20080155077 James Jun 2008 A1
20080176655 James et al. Jul 2008 A1
20080214360 Stirling et al. Sep 2008 A1
20080214903 Orbach Sep 2008 A1
20080249836 Angell et al. Oct 2008 A1
20080275309 Stivoric et al. Nov 2008 A1
20080287751 Stivoric et al. Nov 2008 A1
20090012433 Femstrom et al. Jan 2009 A1
20090018797 Kasama et al. Jan 2009 A1
20090043531 Kahn et al. Feb 2009 A1
20090047645 Dibenedetto et al. Feb 2009 A1
20090048044 Oleson et al. Feb 2009 A1
20090054737 Magar et al. Feb 2009 A1
20090054751 Babashan et al. Feb 2009 A1
20090063193 Barton et al. Mar 2009 A1
20090063293 Mirrashidi et al. Mar 2009 A1
20090093341 James et al. Apr 2009 A1
20090098821 Shinya Apr 2009 A1
20090132197 Rubin et al. May 2009 A1
20090143655 Shani Jun 2009 A1
20090144456 Gelf et al. Jun 2009 A1
20090144639 Nims et al. Jun 2009 A1
20090150178 Sutton et al. Jun 2009 A1
20090156172 Chan Jun 2009 A1
20090163783 Mannheimer et al. Jun 2009 A1
20090171788 Tropper et al. Jul 2009 A1
20090195350 Tsern et al. Aug 2009 A1
20090216499 Tobola et al. Aug 2009 A1
20090262088 Moll-Carrillo et al. Oct 2009 A1
20090264713 Van Loenen et al. Oct 2009 A1
20090271147 Sugai Oct 2009 A1
20090287921 Zhu et al. Nov 2009 A1
20090292332 Li et al. Nov 2009 A1
20090307517 Fehr et al. Dec 2009 A1
20090309742 Alexander et al. Dec 2009 A1
20090318779 Tran Dec 2009 A1
20100023348 Hardee et al. Jan 2010 A1
20100026995 Merritt et al. Feb 2010 A1
20100059561 Ellis et al. Mar 2010 A1
20100063365 Pisani et al. Mar 2010 A1
20100069203 Kawaguchi et al. Mar 2010 A1
20100079291 Kroll et al. Apr 2010 A1
20100106044 Linderman Apr 2010 A1
20100113948 Yang et al. May 2010 A1
20100125729 Baentsch et al. May 2010 A1
20100130873 Yuen et al. May 2010 A1
20100152600 Droitcour et al. Jun 2010 A1
20100158494 King Jun 2010 A1
20100159709 Kotani et al. Jun 2010 A1
20100167783 Alameh et al. Jul 2010 A1
20100179411 Holmstrom et al. Jul 2010 A1
20100185064 Bandic et al. Jul 2010 A1
20100204550 Heneghan et al. Aug 2010 A1
20100205541 Rapaport et al. Aug 2010 A1
20100217099 LeBoeuf et al. Aug 2010 A1
20100222179 Temple et al. Sep 2010 A1
20100240972 Neal Sep 2010 A1
20100249633 Droitcour et al. Sep 2010 A1
20100261987 Kamath et al. Oct 2010 A1
20100274100 Behar et al. Oct 2010 A1
20100292050 DiBenedetto Nov 2010 A1
20100292568 Droitcour et al. Nov 2010 A1
20100292600 DiBenedetto et al. Nov 2010 A1
20100295684 Hsieh et al. Nov 2010 A1
20100298650 Moon et al. Nov 2010 A1
20100298651 Moon et al. Nov 2010 A1
20100298656 Mccombie et al. Nov 2010 A1
20100298661 McCombie et al. Nov 2010 A1
20100311544 Robinette et al. Dec 2010 A1
20100324384 Moon et al. Dec 2010 A1
20100331145 Lakovic et al. Dec 2010 A1
20100331657 Mensinger et al. Dec 2010 A1
20110003665 Burton et al. Jan 2011 A1
20110009051 Khedouri et al. Jan 2011 A1
20110009727 Mensinger et al. Jan 2011 A1
20110022349 Stirling et al. Jan 2011 A1
20110032105 Hoffman et al. Feb 2011 A1
20110066009 Moon et al. Mar 2011 A1
20110066010 Moon et al. Mar 2011 A1
20110066044 Moon et al. Mar 2011 A1
20110080349 Holbein et al. Apr 2011 A1
20110087076 Brynelsen et al. Apr 2011 A1
20110106449 Chowdhary et al. May 2011 A1
20110112442 Meger et al. May 2011 A1
20110118621 Chu May 2011 A1
20110131005 Ueshima et al. Jun 2011 A1
20110153773 Vandwalle Jun 2011 A1
20110167262 Ross et al. Jul 2011 A1
20110184248 Furuta Jul 2011 A1
20110193704 Harper et al. Aug 2011 A1
20110197157 Hoffman et al. Aug 2011 A1
20110214030 Greenberg et al. Sep 2011 A1
20110224508 Moon Sep 2011 A1
20110230729 Shirasaki et al. Sep 2011 A1
20110237911 Lamego et al. Sep 2011 A1
20110237912 Couronne et al. Sep 2011 A1
20110263950 Larson et al. Oct 2011 A1
20110276304 Yin et al. Nov 2011 A1
20120035487 Werner et al. Feb 2012 A1
20120072165 Jallon Mar 2012 A1
20120083705 Yuen et al. Apr 2012 A1
20120083714 Yuen et al. Apr 2012 A1
20120083715 Yuen et al. Apr 2012 A1
20120083716 Yuen et al. Apr 2012 A1
20120084053 Yuen et al. Apr 2012 A1
20120084054 Yuen et al. Apr 2012 A1
20120092157 Tran Apr 2012 A1
20120094649 Porrati et al. Apr 2012 A1
20120116684 Ingrassia, Jr. et al. May 2012 A1
20120119911 Jeon et al. May 2012 A1
20120123232 Najarian et al. May 2012 A1
20120140233 Rockwell et al. Jun 2012 A1
20120143067 Watson et al. Jun 2012 A1
20120150052 Buccheim et al. Jun 2012 A1
20120150074 Yanev et al. Jun 2012 A1
20120165684 Sholder Jun 2012 A1
20120166257 Shiraqarni et al. Jun 2012 A1
20120172733 Park Jul 2012 A1
20120179011 Moon et al. Jul 2012 A1
20120179278 Riley et al. Jul 2012 A1
20120183939 Aragones et al. Jul 2012 A1
20120215328 Schmelzer Aug 2012 A1
20120226471 Yuen et al. Sep 2012 A1
20120226472 Yuen et al. Sep 2012 A1
20120227737 Mastrototaro et al. Sep 2012 A1
20120229270 Morley et al. Sep 2012 A1
20120232432 Kahn et al. Sep 2012 A1
20120245439 Andre et al. Sep 2012 A1
20120253486 Niemimaki Oct 2012 A1
20120255875 Vicente et al. Oct 2012 A1
20120265480 Oshima Oct 2012 A1
20120271121 Della Torre et al. Oct 2012 A1
20120271180 Ren et al. Oct 2012 A1
20120274508 Brown et al. Nov 2012 A1
20120283855 Hoffman et al. Nov 2012 A1
20120296400 Bierman et al. Nov 2012 A1
20120297229 Desai et al. Nov 2012 A1
20120316456 Rahman et al. Dec 2012 A1
20120316471 Rahman et al. Dec 2012 A1
20120324226 Bichsel et al. Dec 2012 A1
20120330109 Tran Dec 2012 A1
20130006718 Nielsen et al. Jan 2013 A1
20130009779 Wittling et al. Jan 2013 A1
20130041590 Burich et al. Feb 2013 A1
20130053661 Alberth et al. Feb 2013 A1
20130072169 Ross et al. Mar 2013 A1
20130073254 Yuen et al. Mar 2013 A1
20130073255 Yuen et al. Mar 2013 A1
20130077823 Mestha et al. Mar 2013 A1
20130077826 Cowperthwaite et al. Mar 2013 A1
20130079607 Gareau et al. Mar 2013 A1
20130080113 Yuen et al. Mar 2013 A1
20130094600 Beziat et al. Apr 2013 A1
20130095459 Tran Apr 2013 A1
20130096843 Yuen et al. Apr 2013 A1
20130102251 Linde et al. Apr 2013 A1
20130106684 Weast et al. May 2013 A1
20130132501 Vandwalle et al. May 2013 A1
20130151196 Yuen et al. Jun 2013 A1
20130158369 Yuen et al. Jun 2013 A1
20130163390 Gossweiler, III et al. Jun 2013 A1
20130166048 Werner et al. Jun 2013 A1
20130173171 Drysdale et al. Jul 2013 A1
20130190903 Balakrishnan et al. Jul 2013 A1
20130191034 Weast et al. Jul 2013 A1
20130203475 Kil et al. Aug 2013 A1
20130209972 Carter et al. Aug 2013 A1
20130211265 Bedingham et al. Aug 2013 A1
20130218053 Kaiser et al. Aug 2013 A1
20130225117 Giacoletto et al. Aug 2013 A1
20130228063 Turner Sep 2013 A1
20130231574 Tran Sep 2013 A1
20130238287 Hoffman et al. Sep 2013 A1
20130245436 Tupin, Jr. et al. Sep 2013 A1
20130261415 Ashe et al. Oct 2013 A1
20130261475 Mochizuki Oct 2013 A1
20130267249 Rosenberg Oct 2013 A1
20130268199 Nielsen et al. Oct 2013 A1
20130268236 Yuen et al. Oct 2013 A1
20130268687 Schrecker Oct 2013 A1
20130274904 Coza et al. Oct 2013 A1
20130280682 Levine et al. Oct 2013 A1
20130281110 Zelinka Oct 2013 A1
20130289366 Chua et al. Oct 2013 A1
20130296666 Kumar et al. Nov 2013 A1
20130296672 O'Neil et al. Nov 2013 A1
20130296673 Thaveeprungsriporn et al. Nov 2013 A1
20130296714 Kassim Nov 2013 A1
20130310896 Mass Nov 2013 A1
20130325396 Yuen et al. Dec 2013 A1
20130331058 Harvey Dec 2013 A1
20130337974 Yanev et al. Dec 2013 A1
20140035761 Burton et al. Feb 2014 A1
20140039284 Niwayama et al. Feb 2014 A1
20140039804 Park et al. Feb 2014 A1
20140039840 Yuen et al. Feb 2014 A1
20140039841 Yuen et al. Feb 2014 A1
20140052280 Yuen et al. Feb 2014 A1
20140067278 Yuen et al. Mar 2014 A1
20140073486 Ahmed Mar 2014 A1
20140074431 Modi Mar 2014 A1
20140077673 Garg et al. Mar 2014 A1
20140088385 Moon et al. Mar 2014 A1
20140094941 Ellis et al. Apr 2014 A1
20140099614 Hu et al. Apr 2014 A1
20140107493 Yuen et al. Apr 2014 A1
20140125618 Panther et al. May 2014 A1
20140135612 Yuen et al. May 2014 A1
20140135631 Brumback et al. May 2014 A1
20140142403 Brumback et al. May 2014 A1
20140164611 Molettiere et al. Jun 2014 A1
20140180022 Stivoric et al. Jun 2014 A1
20140207264 Quy Jul 2014 A1
20140213858 Presura et al. Jul 2014 A1
20140228649 Rayner Aug 2014 A1
20140241626 Sull et al. Aug 2014 A1
20140275821 Beckman Sep 2014 A1
20140275852 Hong et al. Sep 2014 A1
20140275854 Venkatraman et al. Sep 2014 A1
20140275885 Isaacson et al. Sep 2014 A1
20140276119 Venkatraman et al. Sep 2014 A1
20140278139 Hong et al. Sep 2014 A1
20140278229 Hong et al. Sep 2014 A1
20140288390 Hong et al. Sep 2014 A1
20140288391 Hong et al. Sep 2014 A1
20140288392 Hong et al. Sep 2014 A1
20140288435 Richards et al. Sep 2014 A1
20140288436 Venkatraman et al. Sep 2014 A1
20140288438 Venkatraman et al. Sep 2014 A1
20140303523 Hong et al. Oct 2014 A1
20140316305 Venkatraman et al. Oct 2014 A1
20140323827 Ahmed et al. Oct 2014 A1
20140323828 Ahmed et al. Oct 2014 A1
20140343372 Ahmed et al. Nov 2014 A1
20140350356 Ahmed et al. Nov 2014 A1
20140378786 Hong et al. Dec 2014 A1
20140378787 Brumback et al. Dec 2014 A1
20140378844 Fei Dec 2014 A1
20140378872 Hong et al. Dec 2014 A1
20150025393 Hong et al. Jan 2015 A1
20150025394 Hong et al. Jan 2015 A1
20150026647 Park et al. Jan 2015 A1
20150173628 Yuen et al. Jun 2015 A1
20150173631 Richards et al. Jun 2015 A1
20150196256 Venkatraman et al. Jul 2015 A1
20150201853 Hong et al. Jul 2015 A1
20150201854 Hong et al. Jul 2015 A1
20150223708 Richards et al. Aug 2015 A1
20150230743 Silveira et al. Aug 2015 A1
20150230761 Brumback et al. Aug 2015 A1
20150282713 Fei Oct 2015 A1
20150351646 Cervini Dec 2015 A1
20150366469 Harris et al. Dec 2015 A1
20150366504 Connor et al. Dec 2015 A1
20160029968 Lerner et al. Feb 2016 A1
20160034634 Hong et al. Feb 2016 A9
20160058309 Han Mar 2016 A1
20160058312 Han et al. Mar 2016 A1
20160113585 Uedaira et al. Apr 2016 A1
20160183818 Richards et al. Jun 2016 A1
20160302706 Richards et al. Oct 2016 A1
20160345881 Sarantos et al. Dec 2016 A1
20160374567 Breslow et al. Dec 2016 A1
20160374569 Breslow et al. Dec 2016 A1
20170020659 Hyde et al. Jan 2017 A1
20170027523 Venkatraman et al. Feb 2017 A1
20170164848 Nadeau et al. Jun 2017 A1
20170311825 Weekly et al. Nov 2017 A1
20180108802 Chen Apr 2018 A1
20180310846 Lin Nov 2018 A1
20190082985 Hong et al. Mar 2019 A1
20190385708 Hong et al. Dec 2019 A1
20200138309 Weekly et al. May 2020 A1
Foreign Referenced Citations (37)
Number Date Country
1623175 Jun 2005 CN
1729933 Aug 2006 CN
101039617 Sep 2007 CN
100362963 Jan 2008 CN
101615098 Dec 2009 CN
101730503 Jun 2010 CN
101742981 Jun 2010 CN
101940476 Jan 2011 CN
102008811 Apr 2011 CN
202069586 Dec 2011 CN
102389313 Mar 2012 CN
102551686 Jul 2012 CN
102750015 Oct 2012 CN
102781310 Nov 2012 CN
103093420 May 2013 CN
104379055 Feb 2015 CN
1 297 784 Apr 2003 EP
1 586 353 Oct 2005 EP
1 721 237 Aug 2012 EP
11-347021 Dec 1999 JP
2010-169410 May 2010 JP
2178588 Jan 2002 RU
WO 02011019 Feb 2002 WO
WO 2006044677 Apr 2006 WO
WO 06055125 May 2006 WO
WO 06090197 Aug 2006 WO
WO 08038141 Apr 2008 WO
WO 09042965 Apr 2009 WO
WO 12061438 May 2012 WO
WO 12170586 Dec 2012 WO
WO 12170924 Dec 2012 WO
WO 12171032 Dec 2012 WO
WO 2014091424 Jun 2014 WO
WO 2014091424 Jun 2014 WO
WO 15127067 Aug 2015 WO
WO 16003269 Jan 2016 WO
WO 2017190051 Nov 2017 WO
Non-Patent Literature Citations (214)
Entry
“Fitbit automatically tracks your fitness and sleep,” Fitbit Inc., published online at www.fitbit.com, downloaded Sep. 10, 2008, 1 page.
“A Hybrid Discriminative/Generative Approach for Modeling Human Activities”, Lester, et al., Proc. of the Int'l Joint Conf. Artificial intelligence, 2005, pp. 766-772.
“Activity Classification Using Realistic Data From Wearable Sensors”, Parkka, et al, IEEE Transactions on Information Technology in Biomedicine, vol. 10, No. 1, Jan. 2006, pp. 119-128.
“Altimeter and Barometer System”, Clifford, et al., Freescale Semiconductor Aplication Note AN1979, Rev. 3, Nov. 2006.
“An Intelligent Multi-Sensor system for Pedestrian Navigation”, Retscher, Journal of Global Positioning Systems, vol. 5, No. 1, 2006, pp. 110-118.
“Automatic classification of ambulatory movements and evaluation of energy consumptions utilizing accelerometers and barometer”, Ohtaki, et al, Microsystem Technologies, vol. 11, No. 8-10, Aug. 2005, pp. 1034-1040.
“Classification of Human Moving Patterns Using Air Pressure and Acceleration”, Sagawa, et al, Proceedings of the 24th Annual Conference of the IEEE Industrial Electronics Society, vol. 2, Aug.-Sep. 1998, pp. 1214-1219.
“Design of a Wireless Assisted Pedestrian Dead Reckoning System—The NavMote Experience”, Fang, et al, IEEE Transactions on instrumentation and Measurement, vol. 54, No. 6, Dec. 2005, pp. 2342-2358.
“Direct Measurement of Human Movement by Accelerometry”, Godfrey, et al., Medical Engineering & Physics, vol. 30, 2008, pp. 1364-1386.
“Drift-free dynamic height sensor using MEMS IMU aided by MEMS pressure sensor”, Tanigawa, et al, Workshop on Positioning, Navigation and Communication, Mar. 2008, pp. 191-196.
“Evaluation of a New Method of Heading Estimation of Pedestrian Dead Reckoning Using Shoe Mounted Sensors”, Stirling et al., Journal of Navigation, vol. 58, 2005, pp. 31-45.
“Foot Mounted Inertia System for Pedestrian Naviation”, Godha et al., Measurement Science and Technology, vol. 19, No. 7, May 2008, pp. 1-9.
“Improvement of Walking Speed Prediction by Accelerometry and Altimetry, Validated by Satellite Positioning”, Perrin, et al, Medical & Biological Engineering & Computing, vol. 38, 2000, pp. 164-168.
“Indoor Navigation with MEMS Sensors”, Lammel, et al., Proceedings of the Eurosensors XIII conference, vol. 1, No. 1, Sep. 2009, pp. 532-535.
“Non-restricted measurement of walking distance”, Sagawa, et al, IEEE Int'l Conf. on Systems, Man, and Cybernetics, vol. 3, Oct. 2000, pp. 1847-1852.
“On Foot Navigation: When GPS alone is not Enough”, Ladetto, et al, Journal of Navigation, vol. 53, No. 2, Sep. 2000, pp. 279-285.
“SCP 1000-D01/D11 Pressure Sensor as Barometer and Altimeter”, VTI Technologies Application, Jun. 2006, Note 33.
“Suunto LUMI User Guide”, Jun. and Sep. 1997.
“Using MS5534 for altimeters and barometers”, Intersema App., Note AN501, Jan. 2006.
“Validated caloric expenditure estimation using a single body-worn sensor”, Lester, et al, Proc. of the Int'l Conf. on Ubiquitous Computing, 2009, pp. 225-234.
Deepak et al., Plug-and-Play, Single-Chip Photoplethysmography, 34th Annual International Conference of the IEEE EMBS, San Diego, California USA, Aug. 28-Sep. 1, 2012, 4 pages. Jun. 12, 2014 9:37:30 AM PDT No No.
Specification of the Bluetooth™ System, Core Package, version 4.1, Dec. 2013, vols. 0 & 1, 282 pages.
International Search Report dated Aug. 15, 2008, in related application PCT/IB07/03617.
U.S. Office Action, dated Aug. 4, 2014, issued in U.S. Appl. No. 13/924,784.
U.S. Notice of Allowance, dated Nov. 19, 2014, issued in U.S. Appl. No. 13/924,784.
U.S. Office Action, dated Oct. 22, 2014, issued in U.S. Appl. No. 14/290,884.
U.S. Notice of Allowance, dated Feb. 6, 2015, issued in U.S. Appl. No. 14/290,884.
U.S. Office Action, dated Jun. 22, 2015, issued in U.S. Appl. No. 14/693,710.
U.S. Notice of Allowance, dated Jul. 27, 2015, issued in U.S. Appl. No. 14/693,710.
U.S. Notice of Allowance, dated Apr. 15, 2016, issued in U.S. Appl. No. 14/954,753.
U.S. Office Action, dated Oct. 26, 2016, issued in U.S. Appl. No. 15/195,911.
U.S. Notice of Allowance, dated Jan. 23, 2017, issued in U.S. Appl. No. 15/195,911.
U.S. Notice of Allowance, dated Sep. 23, 2014, issued in U.S. Appl. No. 14/292,669.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Oct. 14, 2014, issued in U.S. Appl. No. 14/292,669.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Dec. 31, 2014, issued in U.S. Appl. No. 14/292,669.
U.S. Notice of Allowance, dated Oct. 14, 2014, issued in U.S. Appl. No. 14/295,144.
U.S. Notice of Allowance, dated Dec. 3, 2014, issued in U.S. Appl. No. 14/295,144.
U.S. Notice of Allowance, dated Sep. 26, 2014, issued in U.S. Appl. No. 14/295,158.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Dec. 31, 2014, issued in U.S. Appl. No. 14/295,158.
U.S. Office Action, dated Jan. 23, 2015, issued in U.S. Appl. No. 14/507,184.
U.S. Final Office Action, dated May 11, 2015, issuedin U.S. Appl. No. 14/507,184.
U.S. Notice of Allowance, dated Aug. 11, 2015, issued in U.S. Appl. No. 14/507,184.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Dec. 18, 2015, issued in U.S. Appl. No. 14/507,184.
U.S. Office Action, dated Jan. 26, 2015, issued in U.S. Appl. No. 14/295,161.
U.S. Notice of Allowance, dated Apr. 14, 2015, issued in U.S. Appl. No. 14/295,161.
U.S. Notice of Allowance, dated Jul. 28, 2015, issued in U.S. Appl. No. 14/295,161.
U.S. Office Action, dated May 11, 2015, issued in U.S. Appl. No. 14/673,630.
U.S. Notice of Allowance, dated Nov. 25, 2015, issued in U.S. Appl. No. 14/673,630.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Mar. 21, 2016, issued in U.S. Appl. No. 14/673,630.
U.S. Office Action, dated Jan. 27, 2015, issued in U.S. Appl. No. 14/507,173.
U.S. Notice of Allowance, dated Apr. 17, 2015, issued in U.S. Appl. No. 14/507,173.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Jul. 16, 2015, issued in U.S. Appl. No. 14/507,173.
U.S. Office Action, dated Jun. 8, 2015, issued in U.S. Appl. No. 14/673,634.
U.S. Final Office Action, dated Nov. 4, 2015, issued in U.S. Appl. No. 14/673,634.
U.S. Office Action, dated Jul. 13, 2016, issued in U.S. Appl. No. 14/673,634.
U.S. Office Action, dated Feb. 9, 2017, issued in U.S. Appl. No. 14/673,634.
U.S. Final Office Action, dated Aug. 9, 2017, issued in U.S. Appl. No. 14/673,634.
U.S. Office Action, dated Mar. 27, 2018, issued in U.S. Appl. No. 14/673,634.
U.S. Office Action, dated Aug. 5, 2014, issued in U.S. Appl. No. 14/292,673.
U.S. Notice of Allowance, dated Dec. 8, 2014, issued in U.S. Appl. No. 14/292,673.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Mar. 5, 2015, issued in U.S. Appl. No. 14/292,673.
U.S. Office Action, dated Sep. 18, 2014, issued in U.S. Appl. No. 14/295,059.
U.S. Notice of Allowance, dated Jan. 28, 2015, issued in U.S. Appl. No. 14/295,059.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Mar. 11, 2015, issued in U.S. Appl. No. 14/295,059.
U.S. Office Action, dated Dec. 24, 2014, issued in U.S. Appl. No. 14/295,076.
U.S. Final Office Action, dated Apr. 15, 2015, issued in U.S. Appl. No. 14/295,076.
U.S. Office Action, dated Oct. 22, 2015, issued in U.S. Appl. No. 14/295,076.
U.S. Notice of Allowance, dated May 24, 2016, issued in U.S. Appl. No. 14/295,076.
U.S. Office Action, dated Jan. 12, 2018, issued in U.S. Appl. No. 15/246,387.
U.S. Notice of Allowance, dated Aug. 29, 2018, issued in U.S. Appl. No. 15/246,387.
U.S. Office Action, dated Jul. 31, 2014, issued in U.S. Appl. No. 14/295,122.
U.S. Notice of Allowance, dated Nov. 24, 2014, issued in U.S. Appl. No. 14/295,122.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Jan. 5, 2015, issued in U.S. Appl. No. 14/295,122.
U.S. Office Action dated Dec. 22, 2016, issued in U.S. Appl. No. 14/599,039.
U.S. Final Office Action dated Aug. 3, 2017, issued in U.S. Appl. No. 14/599,039.
U.S. Office Action, dated Mar. 14, 2014, issued in U.S. Appl. No. 14/154,009.
U.S. Office Action, dated Sep. 29, 2014, issued in U.S. Appl. No. 14/154,009.
U.S. Notice of Allowance, dated Jan. 21, 2015, issued in U.S. Appl. No. 14/154,009.
U.S. Office Action, dated Nov. 25, 2014, issued in U.S. Appl. No. 14/154,019.
U.S. Notice of Allowance, dated Mar. 20, 2015, issued in U.S. Appl. No. 14/154,019.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated May 14, 2015, issued in U.S. Appl. No. 14/154,019.
U.S. Office Action, dated Jul. 24, 2018, issued in U.S. Appl. No. 14/696,256.
U.S. Final Office Action, dated Feb. 26, 2019, issued in U.S. Appl. No. 14/696,256.
U.S. Office Action, dated Feb. 19, 2020, issued in U.S. Appl. No. 14/696,256.
U.S. Office Action, dated Dec. 10, 2014, issued in U.S. Appl. No. 14/484,104.
U.S. Notice of Allowance, dated Mar. 19, 2015, issued in U.S. Appl. No. 14/484,104.
U.S. Notice of Allowance (Corrected Notice of Allowability), dated May 6, 2015, issued in U.S. Appl. No. 14/484,104.
U.S. Office Action, dated Dec. 4, 2014, issued in U.S. Appl. No. 14/216,743.
U.S. Final Office Action, dated Apr. 8, 2015, issued in U.S. Appl. No. 14/216,743.
U.S. Office Action, dated Oct. 2, 2015, issued in U.S. Appl. No. 14/216,743.
U.S. Final Office Action, dated Feb. 8, 2016, issued in U.S. Appl. No. 14/216,743.
U.S. Office Action, dated May 16, 2016, issued in U.S. Appl. No. 14/216,743.
U.S. Office Action, dated Jan. 13, 2017, issued in U.S. Appl. No. 14/216,743.
U.S. Examiner's Answer to Appeal Brief before the Patent Trial and Appeal Board [in response to the appeal brief filed Sep. 12, 2017 appealing from the Office action dated Jan. 3, 2017], dated Nov. 30, 2017, issued in U.S. Appl. No. 14/216,743.
U.S. Patent Trial and Appeal Board's Decision on Appeal, dated Oct. 9, 2018, issued in U.S. Appl. No. 14/216,743.
U.S. Notice of Allowance, dated Dec. 17, 2018, issued in U.S. Appl. No. 14/216,743.
U.S. Office Action, dated Mar. 12, 2015, issued in U.S. Appl. No. 14/481,020.
U.S. Final Office Action, dated Jul. 7, 2015, issued in U.S. Appl. No. 14/481,020.
U.S. Office Action, dated Oct. 27, 2015, issued in U.S. Appl. No. 14/481,020.
U.S. Final Office Action, dated May 13, 2016, issued in U.S. Appl. No. 14/481,020.
U.S. Examiner's Answer to Appeal Brief before the Patent Trial and Appeal Board [in response to the appeal brief filed Dec. 9, 2016 appealing from the Office action dated May 13, 2016], dated Jan. 23, 2017, issued in U.S. Appl. No. 14/481,020.
U.S. Patent Trial and Appeal Board's Decision on Appeal, dated Sep. 14, 2018, issued in U.S. Appl. No. 14/481,020.
U.S. Notice of Allowance, dated Nov. 29, 2018, issued in U.S. Appl. No. 14/481,020.
U.S. Office Action, dated Aug. 22, 2014, issued in U.S. Appl. No. 14/250,256.
U.S. Final Office Action, dated Nov. 21, 2014, issued in U.S. Appl. No. 14/250,256.
U.S. Office Action, dated Jul. 8, 2015, issued in U.S. Appl. No. 14/250,256.
U.S. Final Office Action, dated Oct. 23, 2015, issued in U.S. Appl. No. 14/250,256.
U.S. Office Action, dated Mar. 17, 2016, issued in U.S. Appl. No. 14/250,256.
U.S. Final Office Action, dated Jun. 29, 2016, issued in U.S. Appl. No. 14/250,256.
U.S. Office Action, dated Jan. 9, 2017, issued in U.S. Appl. No. 14/250,256.
U.S. Examiner's Answer to the Appeal Brief before the Patent Trial and Appeal Board [in response to the appeal brief filed Jul. 11, 2017 appealing from the Office action dated Jan. 9, 2017], dated Aug. 24, 2017, issued in U.S. Appl. No. 14/250,256.
U.S. Patent Trial and Appeal Board's Decision on Appeal, dated Oct. 9, 2018, issued in U.S. Appl. No. 14/250,256.
U.S. Notice of Allowance, dated Mar. 29, 2019, issued in U.S. Appl. No. 14/250,256.
U.S. Office Action, dated Oct. 7, 2014, issued in U.S. Appl. No. 14/481,762.
U.S. Final Office Action, dated Dec. 19, 2014, issued in U.S. Appl. No. 14/481,762.
U.S. Office Action, dated Jul. 7, 2015, issued in U.S. Appl. No. 14/481,762.
U.S. Final Office Action, dated Nov. 5, 2015, issued in U.S. Appl. No. 14/481,762.
U.S. Office Action, dated May 11, 2016, issued in U.S. Appl. No. 14/481,762.
U.S. Final Office Action, dated Oct. 19, 2016, issued in U.S. Appl. No. 14/481,762.
U.S. Office Action, dated Apr. 12, 2017, issued in U.S. Appl. No. 14/481,762.
U.S. Office Action, dated Nov. 19, 2015, issued in U.S. Appl. No. 14/724,750.
U.S. Notice of Allowance, dated Mar. 8, 2016, issued in U.S. Appl. No. 14/724,750.
U.S. Office Action dated Sep. 8, 2016, issued in U.S. Appl. No. 15/192,447.
U.S. Final Office Action dated Feb. 7, 2017, issued in U.S. Appl. No. 15/192,447.
U.S. Notice of Allowance dated May 24, 2017, issued in U.S. Appl. No. 15/192,447.
U.S. Office Action dated Mar. 15, 2017, issued in U.S. Appl. No. 15/370,303.
U.S. Final Office Action dated Aug. 1, 2017, issued in U.S. Appl. No. 15/370,303.
U.S. Office Action dated Jan. 11, 2018, issued in U.S. Appl. No. 15/370,303.
U.S. Final Office Action dated Jul. 25, 2018, issued in U.S. Appl. No. 15/370,303.
U.S. Office Action dated May 24, 2019, issued in U.S. Appl. No. 15/370,303.
U.S. Office Action, dated Oct. 7, 2014, issued in U.S. Appl. No. 14/292,844.
U.S. Notice of Allowance, dated Feb. 9, 2015, issued in U.S. Appl. No. 14/292,844.
U.S. Office Action, dated Jul. 6, 2015, issued in U.S. Appl. No. 14/640,281.
U.S. Final Office Action, dated Nov. 12, 2015, issued in U.S. Appl. No. 14/640,281.
U.S. Office Action, dated Oct. 6, 2016, issued in U.S. Appl. No. 14/640,281.
U.S. Final Office Action, dated May 4, 2017, issued in U.S. Appl. No. 14/640,281.
U.S. Office Action, dated Jun. 29, 2018, issued in U.S. Appl. No. 14/640,281.
U.S. Final Office Action, dated Feb. 21, 2019, issued in U.S. Appl. No. 14/640,281.
U.S. Notice of Allowance, dated Aug. 2, 2019, issued in U.S. Appl. No. 14/640,281.
U.S. Office Action, dated May 30, 2019, issued in U.S. Appl. No. 15/376,542.
U.S. Notice of Allowance, dated Sep. 20, 2019, issued in U.S. Appl. No. 15/376,542.
U.S. Office Action, dated Mar. 11, 2019, issued in U.S. Appl. No. 15/582,240.
U.S. Notice of Allowance, dated Jun. 14, 2019, issued in U.S. Appl. No. 15/582,240.
U.S. Office Action, dated Apr. 1, 2020, issued in U.S. Appl. No. 15/948,970.
Chinese First Office Action dated Sep. 27, 2016 issued in Application No. CN 201410018701.8.
Chinese Second Office Action dated Jun. 13, 2017 issued in Application No. CN 201410018701.8.
Chinese First Office Action dated Aug. 7, 2015 issued in Application No. CN 201410243180.6.
Chinese First Office Action dated Sep. 2, 2016 issued in Application No. CN 201510745382.5.
Chinese Second Office Action dated Mar. 22, 2017 issued in Application No. CN 201510745382.5.
Chinese First Office Action dated Mar. 22, 2018 issued in Application No. CN 201610284612.7.
Chinese Second Office Action dated Nov. 6, 2018 issued in Application No. CN 201610284612.7.
Chinese First Office Action dated Aug. 3, 2016 issued in Application No. CN 201410243169.X.
Chinese Second Office Action dated Mar. 27, 2017 issued in Application No. CN 201410243169.X.
Chinese Third Office Action dated Sep. 28, 2017 issued in Application No. CN 201410243169.X.
Chinese First Office Action dated Sep. 26, 2016 issued in Application No. CN 201410243178.9.
Chinese Second Office Action dated Jun. 15, 2017 issued in Application No. CN 201410243178.9.
Chinese First Office Action dated Mar. 3, 2017 issued in Application No. CN 201610622453.7.
Chinese Second Office Action dated Sep. 19, 2017 issued in Application No. CN 201610622453.7.
Chinese Third Office Action dated Jan. 24, 2018 issued in Application No. CN 201610622453.7.
Chinese Fourth Office Action dated Jun. 1, 2018 issued in Application No. CN 201610622453.7.
Chinese First Office Action dated Jul. 13, 2017 issued in Application No. CN 201610621114.7.
Chinese Second Office Action dated Apr. 9, 2018 issued in Application No. CN 201610621114.7.
Chinese Third Office Action dated Sep. 14, 2018 issued in Application No. CN 201610621114.7.
Chinese First Office Action dated Jan. 14, 2019 issued in Application No. CN 201510117698.X.
Chinese Second Office Action dated Jun. 21, 2019 issued in Application No. CN 201510117698.X.
Chinese First Office Action dated May 13, 2020, issued in Application No. CN 201610377864.4.
Chinese First Office Action dated Jan. 22, 2020, issued in Application No. CN 201780033558.1.
European Extended Search Report dated Oct. 25, 2016 issued in Application No. EP 16 16 8661.3.
European Office Action dated Mar. 19, 2019 issued in Application No. EP 16 16 8661.3.
European Extended Search Report dated Sep. 9, 2019, issued in Application No. EP 17790575.9.
International Search Report and Written Opinion—PCT/US2017/030190—ISA/US—dated Jul. 7, 2017 (Jul. 7, 2017).
Litigation Document—“Complaint For Patent Infringement,” filed Sep. 3, 2015, in U.S. District Court of Delaware (Court Docket No. 1: 15-cv-00775-RGA).
Litigation Document—“Report On the Filing or Determination of an Action Regarding a Patent or Trademark,” filed Sep. 3, 2015, in U.S. District Court of Delaware (Court Docket No. 1: 15-cv-00775-RGA).
Litigation Document—“Complaint For Patent Infringement,” filed Oct. 29, 2015, in U.S. District Court of Delaware (Court Docket No. 1:15-cv-00990-RGA) [Re: U.S. Pat. Nos. 8,868,377, 8,920,332, and 9,089,760].
Litigation Document—“Report On the Filing or Determination of an Action Regarding a Patent or Trademark,” filed Oct. 29, 2015, in U.S. District Court of Delaware (Court Docket No. 1:15-cv-00990-RGA) [Re: U.S. Pat. Nos. 8,868,377, 8,920,332, and 9,089,760].
Litigation Document—“ORDER No. 24: Initial Determination Granting Respondents' Motion for Summary Determination of Invalidity under 35 U.S.C. § 101 with respect to all Three Asserted Patents and Terminating the Investigation in its Entirety,” filed Jul. 19, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof].
Litigation Document—“Respondents' Opposition to Complainant's Petition for Review of the Initial Determination Granting Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Aug. 8, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) (4446833v1/014972) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof].
Litigation Document—“Declaration of Majid Sarrafzadeh in Support of Complainant's Brief in Opposition to Respondents' Motion for Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Jun. 2, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 37-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof] [Exhibit 7].
Litigation Document—“Kiaei Declaration in Support of Complainant's Supplemental Brief Regarding Construction of “Operating the Heart Rate Monitor in a Worn Detection Mode” under 35 U.S.C. § 112(f),” filed Apr. 29, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 37-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof] [Exhibit 8].
Litigation Document—“Memorandum in Support of Respondents' Motion for Summary Deteninination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed May 23, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) (44325007v1/014972) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof].
Litigation Document—“Grimes Declaration in Support of Complainant's Brief in Opposition to Respondents' Motion for Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Jun. 2, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 37-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof] [Exhibit 28].
Litigation Document—“Complainant's Brief in Opposition to Respondents' Motion for Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Jun. 2, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof].
Litigation Document—“Complainant's Petition for Review of the Initial Determination Granting Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Aug. 1, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof].
Litigation Document—“Summary Pursuant to 19 C.F.R. § 210.43(b)(2) of Complainant's Petition for Review of the Initial Determination Granting Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Aug. 1, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof].
Litigation Document—“Notice of Commission Determination to Review an Initial Determination Granting Respondents' Motion for Summary Determination that Certain Asserted Claims are Directed to Ineligible Subject Matter under 35 U.S.C. § 101; and on Review to Remand the Investigation to the Presiding Administrative Law Judge,” issued Sep. 7, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof].
U.S. Appl. No. 61/736,310, filed Dec. 12, 2012, William Ahmed et al., entitled “Fitness Monitoring Systems and Methods Based on Continuous Collection of Physiological Data,” 61pp [Exhibit 4].
U.S. Appl. No. 61/696,525, filed Sep. 4, 2012, William Ahmed et al., entitled “Fitness Monitoring Systems and Methods Based on Continuous Collection of Physiological Data,” 47pp [Exhibit 5].
Gasparrini et al. (2013) “Evaluation and Possible Improvements of the ANT Protocol for Home Heart Monitoring Applications,” IEEE, 978-1-4673-2874-6/13, 7pp [Exhibit 6].
“UP3™, The world's most advanced tracker,” (Oct. 14, 2015) Jawbone, 10pp [Exhibit 12].
“UP4™, A fitness tracker so advanced it pays,” (Oct. 14, 2015) Jawbone, 12pp [Exhibit 13].
“User's Guide, MIO Drive+ Petite,” User's guide and how-to videos available at www.mioglobal.com, 3pp [Exhibit 16].
“SOLO 915, Heart Rate + Calorie Monitor,” (2009) SPORTLINE®, [retrieved on Oct. 15, 2010 at www.sportline.com] 25pp [Exhibit 17].
U.S. Notice of Allowance dated Oct. 14, 2014 issued in U.S. Appl. No. 14/295,144, 5pp [Exhibit 18].
“Health Touch™ Plus User Guide,” (2011) Timex Group USA, Inc., 12pp [Exhibit 18].
Czarnul, Pawel (Jun. 6-8, 2013) “Design of a Distributed System using Mobile Devices and Workflow Management for Measurement and Control of a Smart Home and Health,” Sopot, Poland, IEEE, pp. 184-192, 10pp [Exhibit 19].
Rabinovich, Roberto A., and Louvaris, Zafeiris et al. (Feb. 8, 2013) “Validity of Physical Activity Monitors During Daily Life in Patients With COPD,” ERJ Express, European Respiratory Society, 28pp [Exhibit 24].
Horvath et al. (2007) “The effect of pedometer position and normal gait asymmetry on step count accuracy,” Appl. Physiol. Nutr. Metab., 32:409-415, 8pp [Exhibit 32].
Graser et al. (2007) “Effects of Placement, Attachment, and Weight Classification on Pedometer Accuracy,” Journal of Physical Activity and Health, 4(4):359-369, 13pp [Exhibit 33].
Vyas et al. (2012) “Machine Learning and Sensor Fusion for Estimating Continuous Energy Expenditure,” AI Magazine, pp. 55-61, 13pp [Exhibit 42].
“New Lifestyles, NL-800 Activity Monitor, User's guide & record book,” (2005), New Lifestyles, Inc., 37pp.
“StepWatch Step Activity Monitor, U.S. Pat. No. 5,485,402,” (2001) StepWatch™, Prosthetics Research Study, 7pp.
Litigation Document—“Plaintiffs Original Complaint For Patent Infringement,” filed Jan. 4, 2016, in U.S. District Court for the Eastern District of North Carolina (Court Docket No. 5:16-cv-00002-FL) [Re: U.S Pat. Nos. 8,923,941, 8,886,269, 8,929,965 and 8,989,830], 11 pages.
Cooper, Daniel (Aug. 16, 2013) Withings Pulse review, http://www.engadget.com/2013/08/16/withings-pulse-revew/, 8 pages.
Dunn et al. (2007) “A Novel Ring Shaped Photodiode for Reflectance Pulse Oximetry in Wireless Applications,” IEEE Sensors Conference, pp. 596-599.
Kim, D. et al. A Linear Transformation Approach for Estimating Pulse Arrival Time. Journal of Applied Mathematics. vol. 2012. Jan. 20, 2012. [Retrieve Jun. 19, 2017]. Retrieved from internet: <https://www.emis.de/journals/HOA/JAM/Volume2012/643653.pdf> pp. 1-12.
LIFETRNR, User Manual (2003, specific date unknown), NB new balance®, Implus Footcare, LLC, 3 pages.
Rainmaker, (Jun. 25, 2012, updated Feb. 16, 2013) “Garmin Swim watch In-Depth Review,” [retrieved on Sep. 9, 2013 at http://www.dcrainmaker.com/2012/06/garmin-swim-in-depth-review.html, 38 pp.
“Withings pulse, Quick Installation Guide” (Jul. 24, 2013) Withings Pulse QIG, v 1.3, withings.com/pulse, 16 pages.
Zijlstra, Wiebren, (2004) “Assessment of spatio-temporal parameters during unconstrained walking,” Eur J Appl Physiol, 92:39-44.
U.S. Appl. No. 14/214,655, filed Mar. 14, 2014, Hong et al.
U.S. Appl. No. 15/494,257, filed Apr. 21, 2017, Richards et al.
U.S. Appl. No. 15/948,970, filed Apr. 9, 2018, Nadeau et al.
U.S. Appl. No. 16/592,599, filed Oct. 3, 2019, Weekly et al.
U.S. Appl. No. 16/798,257, filed Feb. 21, 2020, Wu et al.
Related Publications (1)
Number Date Country
20180092551 A1 Apr 2018 US
Provisional Applications (3)
Number Date Country
61924547 Jan 2014 US
61752826 Jan 2013 US
61830600 Jun 2013 US
Continuations (3)
Number Date Country
Parent 14639409 Mar 2015 US
Child 15827970 US
Parent 14302360 Jun 2014 US
Child 14639409 US
Parent 14156381 Jan 2014 US
Child 14302360 US