The present disclosure relates to systems and methods for capturing bodily activity and synchronizing data transfers between a capture device and a client device.
In recent years, the need for health and fitness has grown tremendously. The growth has occurred due to a better understanding of the benefits of good fitness to overall health and wellness. Unfortunately, although today's modern culture has brought about many new technologies, such as the Internet, connected devices and computers, people have become less active. Additionally, many office jobs require people to sit in front of computer screens for long periods of time, which further reduces a person's activity levels. Furthermore, much of today's entertainment options involve viewing multimedia content, computer social networking, and other types of computer involved interfacing. Although such computer activity can be very productive as well as entertaining, such activity tends to reduce a person's overall physical activity.
To provide users concerned with health and fitness a way of measuring or accounting for their activity or lack thereof, fitness trackers are often used. Fitness trackers are used to measure activity, such as walking, motion, running, sleeping, being inactive, bicycling, exercising on an elliptical trainer, and the like. Usually, the data collected by such fitness trackers can be transferred and viewed on a computing device. However, such data is often provided as a basic accumulation of activity data with complicated or confusing interfaces. In addition, updates between a tracker and a client device usually require wired connectors and/or complex syncing schemes.
It is in this context that embodiments described herein arise.
Broadly speaking, the present invention fills these needs by providing a system and method for detecting and measuring a user's heart rate. It should be appreciated that the present invention can be implemented in numerous ways, including as a process, an apparatus, a system, computer readable media, or a device. Several inventive embodiments of the present invention are described below.
One embodiment provides an activity tracking device having a motion sensor and a processor in a housing. The processor configured for processing motion data produced by the motion sensor. A display screen is integrated with the housing to display metrics that quantify the motion data produced by the motion sensor. A light source is integrated within the housing to enable light to be directed out of the housing at a heart rate monitor location on the housing and a light detector is integrated within the housing. The light detector configured to capture an amount of the light that is reflected back to the light detector, at least a first portion of the light reflected back to the light detector is reflected from a blood vessel under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor is also in communication with the light detector to process the reflected light to identify heart beats of the user and produce an indication of a heart rate that can be displayed on the display screen, as an option, in addition to the metrics that quantify the motion data.
The processor can differentiate between a baseline light scattering and reflectance signal detected between the each one of multiple heart beats and a second light scattering and reflectance signal corresponding to at least one heart beat in the blood vessel. The second reflectance signal being less than the baseline reflectance signal, where the blood vessel scatters more of the light during the at least one heart beat than between each one of the multiple heart beats.
The motion sensor can be one of or include an accelerometer, or a global positioning sensor, or a magnetometer, or a gyroscope, or a rotary encoder, or a calorie measurement sensor, or a moisture measurement sensor, or a displacement sensor, or an ultrasonic sensor, or a pedometer, or an altimeter, or a linear motion sensor, or an angular motion sensor, or a multi-axis motion sensor, or a combination of two or more thereof.
The activity tracking device can also include a communication transceiver configured for communicating via at least one a wireless network, an ambient light sensor, an indicator for visually identifying the heart rate monitor location on the housing and at least one infrared (IR) proximity sensor associated with the light source and light detector. The IR proximity sensor can be configured to activate the light source and light detector upon detecting presence of the skin of the user. Detecting presence of the skin of the user can also function to navigate to one or more metrics of the display screen.
The activity tracking device can also include a pressure detecting system configured for detecting a pressure applied to the heart rate monitor location on the housing with the skin of the user during the identification of heart beats. The activity tracking device can output at least one a feedback signal regarding the detected pressure applied to the heart rate monitor location, the feedback indication being indicative of more or less pressure desired to produce the heart rate, the feedback signal including at least one of a visual signal, a graphic signal, a tactile signal, and an audible signal. The pressure detecting system can include at least one of processing of the reflected light to identify one of an excess pressure, an insufficient pressure or an acceptable pressure from the detected heart beats of the user, or a pressure sensor in the button.
The activity tracking device can also include a display of one or both of waveform data or numerical data when the skin of the user is over the heart rate monitor location and the heart beats are being identified over a sampling time period and upon concluding the sampling time period, displaying the heart rate on the display screen. At least one recalled heart rate can be displayed on the display screen. The light detector can be disposed next to the light source.
The heart rate can be calculated based on an algorithm that detects multiple heart beats in the light received in the light detector within a sampling time period, measures a first time interval between a first beat of the detected heart beats and a second beat of the detected heart beats and divides the sample time interval by the first time interval to determine a first estimate of heart beats detected within the sampling time period. The first estimate of heart beats is extrapolated within the sampling time period to a first estimated heart beats per minute and the first estimate heart beats per minute is output to the display screen.
The heart rate can be calculated based on an algorithm that adds at least one beat to the first estimate of heart beats to produce a second estimate of heart beats and subtracts at least one beat from the first estimate of beats to produce a third estimate of heart beats. The first estimate of heart beats, the second estimate of heart beats and the third estimate of heart beats are scored and a highest scoring estimate of heart beats is selected and output to the display screen.
The light source and the reflected light detector can be selected for any suitable wavelength or suitable band of wavelengths of light ranging from between infrared wavelengths through a human visible spectrum to ultraviolet wavelengths. The light source includes at least one of an infrared (IR) light source, wherein the IR light emitted from the light source produces a deadfront at the heart rate monitor location of the housing or a green light source and the heart rate monitor location includes a translucent green window.
The heart rate monitor location includes a cover that enables infrared (IR) light of the light source or any other wavelength of light from the light source to pass while blocking substantially all light in a human visible spectrum. The light source and light detector can additionally function as a proximity sensor to activate the display screen. The heart rate monitor location can include a button. The button can have an infrared (IR) light transmitting structure. The light source and the light detector can be disposed substantially below the button. The button can also function to navigate to one or more metrics of the display screen. The functions to navigate can be enabled while the heart beats are measured. The skin can be of a finger of a user.
Another embodiment provides a method of tracking activity including sensing motion with a motion sensor, the motion sensor including a processor. The processor is configured for processing motion data produced by the motion sensor. Metrics that quantify the motion data produced by the motion sensing can be displayed on a device display screen integrated with a housing of the motion sensor. A directed light is emitted from a light source, the light source being included in the housing at a heart rate monitor location on the housing and an amount of the light that is reflected back to a light detector is captured by the light detector integrated within the housing. At least a first portion of the light reflected back to the light detector is reflected from a blood vessel disposed under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor further being in communication with the light detector to enable processing of the reflected light to identify heart beats of the user and produce an indication of a heart rate. The indication of the heart rate being displayable on the display screen as an option, in addition to the metrics that quantify the motion data.
The method can also include detecting a pressure applied to the heart rate monitor location of the housing with the skin of the user during the identification of heart beats, the pressure being detected by a pressure detecting system included in the housing and outputting a feedback signal regarding the detected pressure applied to the heart rate monitor location, the feedback indication being indicative of more or less pressure desired to produce the heart rate. One or both of waveform data or numerical data can be displayed when the skin of the user is over the heart rate monitor location and the heart beats are being identified over a sampling time period. The heart rate can be displayed on the display screen upon concluding the sampling time period.
Yet another embodiment provides a heart rate monitor including a light source and a light detector disposed to receive light emitted from the light source and reflected from a blood vessel disposed within in a heart rate monitor subject. A display screen and a processor coupled to the light source, the light detector and the display screen, are also included.
Still another embodiment provides an activity tracking device including a housing including a motion sensor and a processor. The processor is configured for processing motion data produced by the motion sensor. A display screen is integrated with the housing to display metrics that quantify the motion data produced by the motion sensor. A light source is also integrated within the housing to enable light to be directed out of the housing at a heart rate monitor location on the housing and a light detector is integrated within the housing. The light detector is configured to capture an amount of the light that is reflected back to the light detector. At least a first portion of the light reflected back to the light detector is reflected from a blood vessel disposed under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor is also in communication with the light detector to enable processing of the reflected light to identify heart beats of the user and produce an indication of a heart rate that can be displayed on the display screen, as an option, in addition to the metrics that quantify the motion data. Other aspects and advantages of the invention will become apparent from the following detailed description, taken in conjunction with the accompanying drawings, illustrating by way of example the principles of the invention.
The present invention will be readily understood by the following detailed description in conjunction with the accompanying drawings.
Several exemplary embodiments for activity tracking devices and methods capable monitoring and displaying both a user's activities and the user's heart rate will now be described. It will be apparent to those skilled in the art that the present invention may be practiced without some or all of the specific details set forth herein.
The user's heart rate can be measured by directing a light of a suitable wavelength into a user's skin and capturing a portion of the light reflected from a user's blood vessel inside the user's body. The reflected light includes data corresponding to the user's heart beats. Using this heart beat data and filtering methods and systems provide a rapid, accurate measurement of the user's heart rate. Thereby allowing the user to monitor both his activity and his heart rate. Motion data from a motion sensor within the activity tracking device can be used to identify false heart beats and provide a more accurate heart rate indication to the user, even while the user is engaged in a rigorous activity.
Embodiments described in the present disclosure provide systems, apparatus, computer readable media, and methods for analyzing tracked activity data and providing navigation screens and interfaces. Some embodiments are directed to providing navigation interfaces for an activity tracking device. The activity tracking device includes sensors for detecting when physical contact occurs onto the activity tracking device and logic for providing a display action to the screen of the activity tracking device. The physical contact, in one embodiment, can be qualified as an input when the physical contact has a particular characteristic that is predefined. The characteristic can be, when the contact is the result of one or more taps, e.g., physical contact to the activity tracking device by a finger or hand of the user, or object held by a user and used to impart the contact.
In other embodiments, the input can be non-physical, such as proximity sensing input. The proximity sensing input can be processed by an infrared proximity sensor, a thermal sensor, etc. The input can also be by way of a button, voice input, gaze detected input, input processed in response to motion or motion profiles, etc.
It should be noted that there are many inventions described and illustrated herein. The present inventions are neither limited to any single aspect nor embodiment thereof, nor to any combinations and/or permutations of such aspects and/or embodiments. Moreover, each of the aspects of the present inventions, and/or embodiments thereof, may be employed alone or in combination with one or more of the other aspects of the present inventions and/or embodiments thereof. For the sake of brevity, many of those permutations and combinations will not be discussed separately herein.
Further, in the course of describing and illustrating the present inventions, various circuitry, architectures, structures, components, functions and/or elements, as well as combinations and/or permutations thereof, are set forth. It should be understood that circuitry, architectures, structures, components, functions and/or elements other than those specifically described and illustrated, are contemplated and are within the scope of the present inventions, as well as combinations and/or permutations thereof.
The environmental sensors 118 may be in the form of motion detecting sensors 118A. In some embodiments, a motion sensor 118A can be one or more of an accelerometer, or a gyroscope, or a rotary encoder, or a calorie measurement sensor, or a heat measurement sensor, or a moisture measurement sensor, or a displacement sensor, or an ultrasonic sensor, or a pedometer, or an altimeter, or a linear motion sensor, or an angular motion sensor, or a multi-axis motion sensor, or a combination thereof.
The biometric sensors 116 can be defined to measure physiological characteristics of the user that is using the activity tracking device 100. The user interface 114 provides a way for communicating with the activity tracking device 100, in response to user interaction 104. The user interaction 104 can be in the form of physical contact (e.g., without limitation, tapping, sliding, rubbing, multiple taps, gestures, etc.). The biometric sensors 116 can be a one or more proximity sensors 184 capable of detecting the user's presence or touch within a predefined distance or proximity. The proximity sensor 184 can be an infrared (IR) proximity sensor associated with the light source 181 and light detector 182, the IR proximity sensor configured to activate the light source and light detector upon detecting presence of the skin of the user.
The light source 181 and the light detector 182 are located near the external surface of the activity tracking device 100 at a heart rate monitor location 183. The heart rate monitor location 183 can include an indicator such as a marking or an image so the user can easily identify the heart rate monitor location 183. The marking or image can be a raised dot or dimple or a depression or an image of the fingerprint or the heart or any other suitable indication of the heart rate monitor location 183. The heart rate monitor location 183 can include a cover that enables infrared (IR) light of the light source 181 to pass while blocking substantially all light in a human visible spectrum. The heart rate monitor location 183 can include the button 126 or be separate from the button. In one embodiment, the button 126 has an infrared (IR) light transmitting structure and the light source 181 and the light detector 182 are disposed below the button, inside the housing. The button 126 can also provide navigation functions to one or more metrics of the display screen 122.
In some embodiments, the user interface 114 is configured to receive user interaction 104 that is in the form of noncontact input. The noncontact input can be by way of one or more proximity sensors 184, button presses, touch sensitive screen inputs, graphical user interface inputs, voice inputs, sound inputs, etc. The activity tracking device 100 can communicate with a client and/or server 112 using the wireless transceiver 110. The wireless transceiver 110 will allow the activity tracking device 100 to communicate using a wireless connection, which is enabled by wireless communication logic. The wireless communication logic can be in the form of a circuit having radio communication capabilities. The radio communication capabilities can be in the form of a Wi-Fi connection, a Bluetooth connection, a low-energy Bluetooth connection, or any other form of wireless tethering or near field communication. In still other embodiments, the activity tracking device 100 can communicate with other computing devices using a wired connection (not shown). As mentioned, the environmental sensors 118 can detect motion of the activity tracking device 100.
The motion can be activity of the user, such as walking, running, stair climbing, etc. The motion can also be in the form of physical contact received on any surface of the activity tracking device 110, so long as the environmental sensors 118 can detect such motion from the physical contact. As will be explained in more detail below, the physical contact may be in the form of a tap or multiple taps by a finger upon the housing of the activity tracking device 100.
In other embodiments, the device components 102 are positioned substantially in a central position of the wrist attachable device, such as under or proximate to a location where a display screen 122 is located. In the illustrated example, the housing 130 also includes a button 126. The button 126 can be pressed to activate the display screen 122, navigate to various metrics displayed on the screen 122, or turn off the screen 122.
As shown in
Some motions will produce and quantify various types of metrics, such as step count, stairs climbed, distance traveled, very active minutes, calories burned, etc. The physical contact logic 142 can include logic that calculates or determines when particular physical contact can qualify as an input. To qualify as an input, the physical contact detected by biometric sensors 116 should have a particular pattern that is identifiable as input. For example, the input may be predefined to be a double tap input, and the physical contact logic 142 can analyze the motion to determine if a double tap indeed occurred in response to analyzing the sensor data produced by sensors 116, 118.
In other embodiments, the physical contact logic can be programmed to determine when particular physical contacts occurred, the time in between the physical contacts, and whether the one or more physical contacts will qualify within predefined motion profiles that would indicate that an input is desired. If physical contact occurs that is not within some predefined profile or pattern, the physical contact logic will not indicate or qualify that physical contact as an input.
The display interface logic 144 is configured to interface with the processor and the physical contact logic to determine when specific metric data will be displayed on the display screen 122 of the activity tracking device 100. The display interface logic 144 can act to turn on the screen, display metric information, display characters or alphanumeric information, display graphical user interface graphics, or combinations thereof. Alarm management logic 146 can function to provide a user interface and settings for managing and receiving input from a user to set an alarm. The alarm management logic can interface with a timekeeping module (e.g., clock, calendar, time zone, etc.), and can trigger the activation of an alarm. The alarm can be in the form of an audible alarm or a non-audible alarm.
A non-audible alarm can provide such alarm by way of a vibration. The vibration can be produced by a motor integrated in the activity tracking device 100. The vibration can be defined to include various vibration patterns, intensities, and custom set patterns. The vibration produced by the motor or motors of the activity tracking device 100 can be managed by the alarm management logic 146 in conjunction with processing by the processor 106. The wireless communication logic 148 is configured for communication of the activity tracking device with another computing device by way of a wireless signal. The wireless signal can be in the form of a radio signal. As noted above, the radio signal can be in the form of a Wi-Fi signal, a Bluetooth signal, a low energy Bluetooth signal, or combinations thereof. The wireless communication logic can interface with the processor 106, storage 108 and battery 154 of device 100, for transferring activity data, which may be in the form of motion data or processed motion data, stored in the storage 108 to the computing device.
In one embodiment, processor 106 functions in conjunction with the various logic components 140, 142, 144, 146, and 148. The processor 106 can, in one embodiment, provide the functionality of any one or all of the logic components. In other embodiments, multiple chips can be used to separate the processing performed by any one of the logic components and the processor 106. Sensors 116, 118 can communicate via a bus with the processor 106 and/or the logic components. The storage 108 is also in communication with the bus for providing storage of the motion data processed or tracked by the activity tracking device 100. Battery 154 is provided for providing power to the activity tracking device 100.
In one embodiment, remote device 200 communicates with activity tracking device 100 over a Bluetooth connection. In one embodiment, the Bluetooth connection is a low energy Bluetooth connection (e.g., Bluetooth LE, BLE, or Bluetooth Smart). Low energy Bluetooth is configured for providing low power consumption relative to standard Bluetooth circuitry. Low energy Bluetooth uses, in one embodiment, a 2.4 GHz radio frequency, which allows for dual mode devices to share a single radio antenna. In one embodiment, low energy Bluetooth connections can function at distances up to 50 meters, with over the air data rates ranging between 1-3 megabits (Mb) per second. In one embodiment, a proximity distance for communication can be defined by the particular wireless link, and is not tied to any specific standard. It should be understood that the proximity distance limitation will change in accordance with changes to existing standards and in view of future standards and/or circuitry and capabilities.
Remote device 200 can also communicate with the Internet 160 using an Internet connection. The Internet connection of the remote device 200 can include cellular connections, wireless connections such as Wi-Fi, and combinations thereof (such as connections to switches between different types of connection links). The remote device, as mentioned above, can be a smartphone or tablet computer, or any other type of computing device having access to the Internet and with capabilities for communicating with the activity tracking device 100.
A server 220 is also provided, which is interfaced with the Internet 160. The server 220 can include a number of applications that service the activity tracking device 100, and the associated users of the activity tracking device 100 by way of user accounts. For example, the server 220 can include an activity management application 224. The activity management application 224 can include logic for providing access to various devices 100, which are associated with user accounts managed by server 220. Server 220 can include storage 226 that includes various user profiles associated with the various user accounts. The user account 228a for user A and the user account 228n for user N are shown to include various information.
The information can include, without limitation, data associated with a display scroll order 230, user data, etc. As will be described in greater detail below, the display scroll order 230 includes information regarding a user's preferences, settings, and configurations which are settable by the user or set by default at the server 220 when accessing a respective user account. The storage 226 will include any number of user profiles, depending on the number of registered users having user accounts for their respective activity tracking devices. It should also be noted that a single user account can have various or multiple devices associated therewith, and the multiple devices can be individually customized, managed and accessed by a user. In one embodiment, the server 220 provides access to a user to view the user data 232 associated with activity tracking device.
The user data 232 viewable by the user includes the tracked motion data, which is processed to identify a plurality of metrics associated with the motion data. The user data 232 viewable by the user can include user heart beat and heart rate data 232A, which is processed to identify a plurality of metrics associated with the user's heart beat.
The metrics are shown in various graphical user interfaces of a website enabled by the server 220. The website can include various pages with graphical user interfaces for rendering and displaying the various metrics for view by the user associated with the user account. In one embodiment, the website can also include interfaces that allow for data entry and configuration by the user.
The configurations can include defining which metrics will be displayed on the activity tracking device 100. In addition, the configurations can include identification of which metrics will be a first metric to be displayed on the activity tracking device. The first metric to be displayed by the activity tracking device can be in response to a user input at the activity tracked device 100. As noted above, the user input can be by way of physical contact. The physical contact is qualified by the processor and/or logic of the activity tracking device 100 to determine if the physical contact should be treated as an input. The input can trigger or cause the display screen of the activity tracking device 100 to be turned on to display a specific metric, that is selected by the user as the first metric to display. In another embodiment, the first metric displayed in response to the input can be predefined by the system as a default.
The configuration provided by the user by way of the server 220 and the activity management application 224 can also be provided by way of the activity tracking application 202 of the computing device 200. For example, the activity tracking application 202 can include a plurality of screens that also display metrics associated with the captured motion data of the activity tracking device 100. The activity tracking application 202 can also allow for user input and configuration at various graphical user interface screens to set and define which input will produce display of the first metric. In other embodiments, in addition to identifying the first metric to be displayed in response to the input, which may be physical contact, the configuration can allow an ordering of which metrics will be displayed in a specific scroll order.
In another embodiment, the scroll order of the metrics is predefined. In some embodiments, the input provided by the user by way of the physical contact can be pre-assigned to a specific metric in the scroll order. For example, the scroll order can remain the same, while the input can allow the screen to jump to a specific entry in the scroll order. Jumping to a specific entry can be viewed as a shortcut to a specific entry that is desired to be seen first by the user upon providing physical contact or input to the device 100.
The heart rate measuring system includes a light source 181 and a reflected light detector 182. The light source 181 and the reflected light detector 182 are located close together in the activity tracking devices 300, 300′, 300″. In one embodiment the light source 181 and the reflected light detector 182 can be immediately adjacent. The light source 181 and the reflected light detector 182 can be included in a single package and/or a single integrated circuit. The light source 181 and the reflected light detector 182 can be selected for any one suitable wavelength or suitable band of wavelengths of light ranging from between infrared, through a human visible spectrum to ultraviolet wavelengths. The heart rate monitor location 183 can include a cover that enables light of the light source to pass while blocking substantially all light in a human visible spectrum. The cover can be a smoked filter or other suitable filter color or shaded plastic or glass or shaded glass, transparent or translucent glass or plastic or ceramic or any other suitable material capable of allowing the desired wavelengths of light to pass through the cover. In one embodiment, the light source 181 uses an infrared (IR) light and the IR light produces a deadfront at the heart rate monitor location 183. Where a deadfront is defined as a continuous surface such that the cover is not easily discernable from the remaining surface of the housing. A deadfront cover is substantially hidden from the user however a light source 181 or the display screen 122 can emits sufficient light to pass through the cover. In another embodiment, the light source 181 can be a green light and the heart rate monitor location 183 can include a translucent green window.
In operation, the user places the skin 306 of a finger tip 302 or other body part over the light source 181. The light source 181 directs the light 310 into the skin 306. The light 310 passes through the skin 306 to a blood vessel 304 such as an artery, vein, or a capillary within the finger 302. A reflected portion 312 of the light 310 is reflected from the blood vessel 304 toward the reflected light detector 182. The light detector 182 outputs a signal corresponding to the reflected portion 312 of the light. The signal is coupled to a processor 106 for processing configured to identify heart beats of the user and produce an indication of a heart rate. The indication of the heart rate can be displayed on the display screen 122.
One embodiment may use a portion of the teachings of detecting heart beats by reflecting light from a blood vessel, as taught, in part by “Plug-and-Play, Single-Chip Photoplethysmography” by Deepak Chandrasekar, et al., pages 3243-3246, presented 34th Annual International Conference of the IEEE EMBS, San Diego, Calif. USA, 28 Aug.-1 Sep. 2012 which is incorporated by reference herein for all purposes.
Chandrasekar, et al, provides in pertinent part “a digital OPS can be used as a high-performance, reflectance-mode PPG sensor . . . LED emits light into the tissue, where it experiences diffuse reflection from the tissue and capillary bed. This establishes a baseline reflectance signal which is detected at the PD. When a pulse wave propagates through the capillary bed, the reflectance signal falls slightly (0.5-5%) due to light scattering. The change is detected by the PD and processed by embedded amplification and signal processing circuitry” (Page 3244, column 1, line 34 through column 2, line 7 and
It should be understood that the teachings described by are Chandrasekar, et al only examples and other examples can include different and additional processes and systems as described in more detail throughout this disclosure. Further, Chandrasekar, et al, fails to teach suitable filtering to provide accurate hear rate indications. Further still, Chandrasekar, et al, cannot discern motion caused false heart beat detections from actual hear beat detections.
The solid line peaks A, B, C, E, G and I correspond to actual detected heart beats. The dashed peak F corresponds to a phantom beat detection. The phantom peak F can be caused by movement of the user's finger or the activity tracking device 300. The phantom peak F can be caused by noise or some other cause.
The dotted peaks D and H correspond to approximate occurrences of heart beats that should have been detected, based on the timing of the actually detected peaks A, B, C, E, G and I. The dotted peaks D and H are not actually detected peaks and are shown in the graphical representation 420 for reference purposes as will be described in more detail below.
The raw heart beat data is next filtered to determine a best guess heart rate.
Human heart beats are typically substantially evenly spaced, thus the seven detected peaks are separated by even time intervals across the sampling time period STP as shown in the estimated 7 peaks detected line. However, the filtering process evaluates the detection process to determine if the initial estimate of seven peaks is accurate as described in
In an operation 510, the light source 181 emits a light 320 into the user's skin and at least a portion of the light 322 is reflected off of a blood vessel 304 inside the user's skin. The detector 182 receives the reflected light 322 in an operation 515 and outputs the raw data of the detected beats within a sampling time period STP, as shown in
As shown in
In an operation 530, at least one peak is subtracted from the initial estimate of seven beats to produce a corresponding at least one subtracted estimate. In the example above two beats are subtracted resulting in corresponding subtracted estimates of a six beat estimate and a five beat estimate. It should be understood that only one or more than two beats could be subtracted. As shown in the graphical representation 450, estimated lines of beats corresponding to five, six, eight and nine peaks are shown evenly distributed across the sampling time period STP.
In an operation 535, each of the five estimated lines of beats are compared to the actually detected peaks A, B, C, E, F, G and I to determine how well each line of beats scores as most closely matching the actually detected peaks A, B, C, E, F, G and I. By way of example, the five beats estimate line closely corresponds to only actually detected peak A and phantom peak F resulting in a corresponding score of 2. Further, the six beats estimate line closely corresponds to only actually detected peaks A, E and G resulting in a corresponding score of 3. Further, the initial estimate of seven beats estimate line closely corresponds to only actually detected peaks A, B and C and phantom peak F resulting in a corresponding score of 4. The eight beats estimate line closely corresponds to peaks A, B, C, E, G and I resulting in a corresponding score of 6. Finally, the nine beats estimate line closely corresponds to peaks A, B, C and I and phantom peak F resulting in a corresponding score of 5.
In an operation 540, the eight beats estimate provides the highest score of 6 and is therefore selected as a potential heart rate for further evaluation. However, only seven peaks were detected by the heart rate monitoring system and thus the eight beats potential heart rate could be an error.
In an operation 545, intervals between the seven detected beats are measured. Recall that human heart beats are substantially evenly spaced and there are three substantially evenly spaced peaks A, B, C in the seven detected beats. The time interval between the substantially evenly spaced peaks A, B, C is selected as a potential beat interval.
In an operation 550, the potential beat interval is compared to the remaining detected beat intervals between peaks C and E, between peaks E and F, between peaks F and G, and between peaks G and I to identify any timing intervals closely corresponding to whole number multiples of the potential beat interval so as to identify missed beats. In the above example, missed peak D is found between detected peaks C and E and missed peak H is found between detected peaks G and I. The timing intervals between peaks E and F and between peaks F and G are substantially less than the potential beat interval and thus peak F is identified as a phantom peak and the phantom peak F is eliminated in an operation 555.
In an operation 560, the estimated heart beats are scored similar to the scoring in operation 535 above. The combination of the actually detected peaks A, B, C, E, G and I and the missed peaks D and H can be evaluated for each of the estimated lines of beats. The eight beats potential heart rate would yield a resulting score of 8, which would be higher than the scores of 3, 2, 3, 4 corresponding to five, six, seven and nine beats estimate lines, respectively. The eight beats potential heart rate can therefore be determined as the best guess in an operation 565.
In an operation 570, the processor 106 calculates the heart rate. The sampling time period STP is extrapolated to 60 seconds and the corresponding heart rate in heart beats per minute is output to the display screen. By way of example, the sampling time period STP is 5.2 seconds in the above example. And 60 seconds divided by 5.2 seconds is equal to 11.54. And further, 11.54 multiplied by the seven time intervals between the evenly spaced estimated eight beats yields a calculated heart rate of 81 beats per minute. In an operation 575, the processor 106 outputs the calculated heart rate to the display screen 122.
It should be noted that the heart rate measuring system can continue to detect, refine and filter the detected heart beats as long as the user's skin is in sufficiently close proximity to the light source 181 and detector 182.
Alternatively, the excessive pressure or insufficient pressure can degrade the quality of the heart beat data in the reflected light 322 to a level that the heart rate cannot be suitably monitored. Similarly, the user's motion, such as running, can interfere with and produce artifacts and false readings. The processor 106 can use the motion sensor to identify and filter out the falsely detected heart beats from the detected heart beats. By way of example, the processor can detect when the user steps down and the vibration through the user's body and identify a detected heart beat coinciding with the step as a suspected false heart beat.
From operation 520 in
The detected heart beats are constantly monitored for usability, in an operation 720. If increasing the pressure provides usable heart beat data then the method operations return to operation 525 in
The detected heart beats are constantly monitored for usability, in an operation 730. If decreasing the pressure provides usable heart beat data then the method operations return to operation 525 in
In one embodiment, the device collects one or more types of physiological and/or environmental data from embedded sensors and/or external devices and communicates or relays such metric information to other devices, including devices capable of serving as Internet-accessible data sources, thus permitting the collected data to be viewed, for example, using a web browser or network-based application. For example, while the user is wearing an activity tracking device, the device may calculate and store the user's step count using one or more sensors. The device then transmits data representative of the user's step count to an account on a web service, computer, mobile phone, or health station where the data may be stored, processed, and visualized by the user. Indeed, the device may measure or calculate a plurality of other physiological metrics in addition to, or in place of, the user's step count.
Some physiological metrics include, but are not limited to, energy expenditure (for example, calorie burn), floors climbed and/or descended, heart rate, heart rate variability, heart rate recovery, location and/or heading (for example, through GPS), elevation, ambulatory speed and/or distance traveled, swimming lap count, bicycle distance and/or speed, blood pressure, blood glucose, skin conduction, skin and/or body temperature, electromyography, electroencephalography, weight, body fat, caloric intake, nutritional intake from food, medication intake, sleep periods (i.e., clock time), sleep phases, sleep quality and/or duration, pH levels, hydration levels, and respiration rate. The device may also measure or calculate metrics related to the environment around the user such as barometric pressure, weather conditions (for example, temperature, humidity, pollen count, air quality, rain/snow conditions, wind speed), light exposure (for example, ambient light, UV light exposure, time and/or duration spent in darkness), noise exposure, radiation exposure, and magnetic field.
Still further, other metrics can include, without limitation, calories burned by a user, weight gained by a user, weight lost by a user, stairs ascended, e.g., climbed, etc., by a user, stairs descended by a user, variation in the user's altitude, steps taken by a user during walking or running, a number of rotations of a bicycle pedal rotated by a user, sedentary activity data, driving a vehicle, a number of golf swings taken by a user, a number of forehands of a sport played by a user, a number of backhands of a sport played by a user, or a combination thereof. In some embodiments, sedentary activity data is referred to herein as inactive activity data or as passive activity data. In some embodiments, when a user is not sedentary and is not sleeping, the user is active. In some embodiments, a user may stand on a monitoring device that determines a physiological parameter of the user. For example, a user stands on a scale that measures a weight, a body fat percentage, a biomass index, or a combination thereof, of the user.
Furthermore, the device or the system collating the data streams may calculate metrics derived from this data. For example, the device or system may calculate the user's stress and/or relaxation levels through a combination of heart rate variability, skin conduction, noise pollution, and sleep quality. In another example, the device or system may determine the efficacy of a medical intervention (for example, medication) through the combination of medication intake, sleep and/or activity data. In yet another example, the device or system may determine the efficacy of an allergy medication through the combination of pollen data, medication intake, sleep and/or activity data. These examples are provided for illustration only and are not intended to be limiting or exhaustive.
This information can be associated to the users account, which can be managed by an activity management application on the server. The activity management application can provide access to the users account and data saved thereon. The activity manager application running on the server can be in the form of a web application. The web application can provide access to a number of websites screens and pages that illustrate information regarding the metrics in various formats. This information can be viewed by the user, and synchronized with a computing device of the user, such as a smart phone.
In one embodiment, the data captured by the activity tracking device 100 is received by the computing device, and the data is synchronized with the activity measured application on the server. In this example, data viewable on the computing device (e.g. smart phone) using an activity tracking application (app) can be synchronized with the data present on the server, and associated with the user's account. In this way, information entered into the activity tracking application on the computing device can be synchronized with application illustrated in the various screens of the activity management application provided by the server on the website.
The user can therefore access the data associated with the user account using any device having access to the Internet. Data received by the network 176 can then be synchronized with the user's various devices, and analytics on the server can provide data analysis to provide recommendations for additional activity, and or improvements in physical health. The process therefore continues where data is captured, analyzed, synchronized, and recommendations are produced. In some embodiments, the captured data can be itemized and partitioned based on the type of activity being performed, and such information can be provided to the user on the website via graphical user interfaces, or by way of the application executed on the users smart phone (by way of graphical user interfaces).
In an embodiment, the sensor or sensors of a device 100 can determine or capture data to determine an amount of movement of the monitoring device over a period of time. The sensors can include, for example, an accelerometer, a magnetometer, a gyroscope, or combinations thereof. Broadly speaking, these sensors are inertial sensors, which capture some movement data, in response to the device 100 being moved. The amount of movement (e.g., motion sensed) may occur when the user is performing an activity of climbing stairs over the time period, walking, running, etc. The monitoring device may be worn on a wrist, carried by a user, worn on clothing (using a clip, or placed in a pocket), attached to a leg or foot, attached to the user's chest, waist, or integrated in an article of clothing such as a shirt, hat, pants, blouse, glasses, and the like. These examples are not limiting to all the possible ways the sensors of the device can be associated with a user or thing being monitored.
In other embodiments, a biological sensor can determine any number of physiological characteristics of a user. As another example, the biological sensor may determine heart rate, a hydration level, body fat, bone density, fingerprint data, sweat rate, and/or a bioimpedance of the user. Examples of the biological sensors include, without limitation, a biometric sensor, a physiological parameter sensor, a pedometer, or a combination thereof.
In some embodiments, data associated with the user's activity can be monitored by the applications on the server and the users device, and activity associated with the user's friends, acquaintances, or social network peers can also be shared, based on the user's authorization. This provides for the ability for friends to compete regarding their fitness, achieve goals, receive badges for achieving goals, get reminders for achieving such goals, rewards or discounts for achieving certain goals, etc.
In some embodiments the heart beats are detected by processing the light received in the light detector, within the sampling time period. A first time interval between a first beat of the detected heart beats and a second heart beat is measured. The sample time interval can be divided by the first time interval to determine a first estimate of heart beats detected within the sampling time period and the first estimate of heart beats within the sampling time period can be extrapolated to a first estimated heart beats per minute. The first estimate heart beats per minute can be output to the display screen. At least one beat can be added to the first estimate of heart beats to produce a second estimate of heart beats that can be scored with the first estimate of heart beats. A highest scoring of the first estimate of heart beats and the second estimate of heart beats can be selected and output to the display screen.
In another embodiment, at least one beat can be subtracted from the first estimate of beats to produce a third estimate of heart beats. The first estimate of heart beats and the third estimate of heart beats can be scored. A highest scoring of the first estimate of heart beats and the third estimate of heart beats can be selected and output to the display screen.
In another embodiment, identifying the heart beats of the user and producing an indication of a heart rate can include identifying and filtering a falsely detected heart beat coinciding with motion detected by the motion sensor. By way of example, a user's motion may be erroneously identified as a heart beat. The processor can compare detected motion (i.e., motion data, instantaneous shocks, etc.) to the detected heart beats and identify heart beats that coincide with motion data. Further, as the motion data and heart beat data are compiled over time, detected motions that often produce corresponding erroneously detected heart beats can be identified and filtered from the detected heart beats.
In another embodiment, activity tracking device includes a housing including a motion sensor and a processor. The processor is configured for processing motion data produced by the motion sensor. A display screen is integrated with the housing to display metrics that quantify the motion data produced by the motion sensor. A light source is integrated within the housing to enable light to be directed out of the housing at a heart rate monitor location on the housing. A light detector is also integrated within the housing. The light detector is configured to capture an amount of the light that is reflected back to the light detector, at least a first portion of the light reflected back to the light detector is reflected from a blood vessel disposed under a skin of a user when the user places the skin over the heart rate monitor location on the housing. The processor can be in communication with the light detector to enable processing of the reflected light to identify heart beats of the user and produce an indication of a heart rate. The indication of the heart rate being displayable on the display screen as an option, in addition to the metrics that quantify the motion data. The heart rate can be calculated based on an algorithm that calculates a first estimate of heart beats per minute corresponding to detected heart beats in the light received in the light detector within a sampling time period. A refined estimate of heart beats per minute can be calculated by adding at least one beat to the first estimate of heart beats to produce a second estimate of heart beats and subtracting at least one beat from the first estimate of beats to produce a third estimate of heart beats. The first estimate of heart beats, the second estimate of heart beats and the third estimate of heart beats are scored and a highest scoring estimate of heart beats is selected and output to the display screen.
As noted, an activity tracking device 100 can communicate with a computing device (e.g., a smartphone, a tablet computer, a desktop computer, or computer device having wireless communication access and/or access to the Internet). The computing device, in turn, can communicate over a network, such as the Internet or an Intranet to provide data synchronization. The network may be a wide area network, a local area network, or a combination thereof. The network may be coupled to one or more servers, one or more virtual machines, or a combination thereof. A server, a virtual machine, a controller of a monitoring device, or a controller of a computing device is sometimes referred to herein as a computing resource. Examples of a controller include a processor and a memory device.
In one embodiment, the processor may be a general purpose processor. In another embodiment, the processor can be a customized processor configured to run specific algorithms or operations. Such processors can include digital signal processors (DSPs), which are designed to execute or interact with specific chips, signals, wires, and perform certain algorithms, processes, state diagrams, feedback, detection, execution, or the like. In some embodiments, a processor can include or be interfaced with an application specific integrated circuit (ASIC), a programmable logic device (PLD), a central processing unit (CPU), or a combination thereof, etc.
In some embodiments, one or more chips, modules, devices, or logic can be defined to execute instructions or logic, which collectively can be viewed or characterized to be a processor. Therefore, it should be understood that a processor does not necessarily have to be one single chip or module, but can be defined from a collection of electronic or connecting components, logic, firmware, code, and combinations thereof.
Examples of a memory device include a random access memory (RAM) and a read-only memory (ROM). A memory device may be a Flash memory, a redundant array of disks (RAID), a hard disk, or a combination thereof.
Embodiments described in the present disclosure may be practiced with various computer system configurations including hand-held devices, microprocessor systems, microprocessor-based or programmable consumer electronics, minicomputers, mainframe computers and the like. Several embodiments described in the present disclosure can also be practiced in distributed computing environments where tasks are performed by remote processing devices that are linked through a wire-based or wireless network.
With the above embodiments in mind, it should be understood that a number of embodiments described in the present disclosure can employ various computer-implemented operations involving data stored in computer systems. These operations are those requiring physical manipulation of physical quantities. Any of the operations described herein that form part of various embodiments described in the present disclosure are useful machine operations. Several embodiments described in the present disclosure also relate to a device or an apparatus for performing these operations. The apparatus can be specially constructed for a purpose, or the apparatus can be a computer selectively activated or configured by a computer program stored in the computer. In particular, various machines can be used with computer programs written in accordance with the teachings herein, or it may be more convenient to construct a more specialized apparatus to perform the required operations.
Various embodiments described in the present disclosure can also be embodied as computer-readable code on a non-transitory computer-readable medium. The computer-readable medium is any data storage device that can store data, which can thereafter be read by a computer system. Examples of the computer-readable medium include hard drives, network attached storage (NAS), ROM, RAM, compact disc-ROMs (CD-ROMs), CD-recordables (CD-Rs), CD-rewritables (RWs), magnetic tapes and other optical and non-optical data storage devices. The computer-readable medium can include computer-readable tangible medium distributed over a network-coupled computer system so that the computer-readable code is stored and executed in a distributed fashion.
Although the method operations were described in a specific order, it should be understood that other housekeeping operations may be performed in between operations, or operations may be performed in an order other than that shown, or operations may be adjusted so that they occur at slightly different times, or may be distributed in a system which allows the occurrence of the processing operations at various intervals associated with the processing.
Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, it will be apparent that certain changes and modifications can be practiced within the scope of the appended claims. Accordingly, the present embodiments are to be considered as illustrative and not restrictive, and the various embodiments described in the present disclosure are not to be limited to the details given herein, but may be modified within the scope and equivalents of the appended claims.
This application is a continuation of and claims priority from U.S. patent application Ser. No. 14/639,409, filed on Mar. 5, 2015 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety and is a continuation of U.S. patent application Ser. No. 14/302,360 (granted as U.S. Pat. No. 9,039,614), filed on Jun. 11, 2014 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety. U.S. patent application Ser. No. 14/302,360 is a continuation of and claims priority from U.S. patent application Ser. No. 14/156,381 (granted as U.S. Pat. No. 8,827,906), filed on Jan. 15, 2014 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety. This application also claims priority, through U.S. Pat. No. 8,827,906, from U.S. Provisional Patent Application No. 61/924,547 filed on Jan. 7, 2014 and entitled “Methods, Systems and Devices for Measuring Fingertip Heart Rate,” which is incorporated herein by reference in its entirety. This application also claims priority, through U.S. Pat. No. 8,827,906, from U.S. Provisional Patent Application No. 61/752,826 filed on Jan. 15, 2013 and entitled “Portable Monitoring Devices and Methods of Operating Same,” which is incorporated herein by reference in its entirety. This application also claims priority through U.S. Pat. No. 8,827,906, from U.S. Provisional Patent Application No. 61/830,600 filed on Jun. 3, 2013 and entitled “Portable Monitoring Devices and Methods of Operating Same,” which is incorporated herein by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
2717736 | Schlesinger | Sep 1955 | A |
2827309 | Fred | Mar 1958 | A |
2883255 | Anderson | Apr 1959 | A |
3163856 | Kirby | Dec 1964 | A |
3250270 | Walter | May 1966 | A |
3522383 | Chang | Jul 1970 | A |
3608545 | Novack et al. | Sep 1971 | A |
3918658 | Beller | Nov 1975 | A |
4192000 | Lipsey | Mar 1980 | A |
4244020 | Ratcliff | Jan 1981 | A |
4258719 | Lewyn | Mar 1981 | A |
4281663 | Pringle | Aug 1981 | A |
4284849 | Anderson et al. | Aug 1981 | A |
4312358 | Barney | Jan 1982 | A |
4367752 | Jimenez et al. | Jan 1983 | A |
4390922 | Pelliccia | Jun 1983 | A |
4407295 | Steuer et al. | Oct 1983 | A |
4425921 | Fujisaki et al. | Jan 1984 | A |
4575804 | Ratcliff | Mar 1986 | A |
4578769 | Frederick | Mar 1986 | A |
4617525 | Lloyd | Oct 1986 | A |
4771792 | Seale | Sep 1988 | A |
4781195 | Martin | Nov 1988 | A |
4846183 | Martin | Jul 1989 | A |
4887249 | Thinesen | Dec 1989 | A |
4930518 | Hrushesky et al. | Jun 1990 | A |
4960126 | Conlon et al. | Oct 1990 | A |
4977509 | Pitchford et al. | Dec 1990 | A |
5036856 | Thornton | Aug 1991 | A |
5058427 | Brandt | Oct 1991 | A |
5101831 | Koyama et al. | Apr 1992 | A |
5224059 | Nitta et al. | Jun 1993 | A |
5295085 | Hoffacker | Mar 1994 | A |
5301154 | Suga | Apr 1994 | A |
5314389 | Dotan | May 1994 | A |
5318597 | Hauck et al. | Jun 1994 | A |
5323650 | Fullen et al. | Jun 1994 | A |
5365930 | Takashima et al. | Nov 1994 | A |
5446705 | Haas et al. | Aug 1995 | A |
5456648 | Edinburg et al. | Oct 1995 | A |
5490523 | Isaacson et al. | Feb 1996 | A |
5513649 | Gevins et al. | May 1996 | A |
5553296 | Forrest et al. | Sep 1996 | A |
5583776 | Levi et al. | Dec 1996 | A |
5645509 | Brewer et al. | Jul 1997 | A |
5671162 | Werbin | Sep 1997 | A |
5704350 | Williams, III | Jan 1998 | A |
5724265 | Hutchings | Mar 1998 | A |
5734625 | Kondo | Mar 1998 | A |
5738104 | Lo et al. | Apr 1998 | A |
5817008 | Ratert et al. | Oct 1998 | A |
5830137 | Scharf | Nov 1998 | A |
5890128 | Diaz et al. | Mar 1999 | A |
5891042 | Sham et al. | Apr 1999 | A |
5894454 | Kondo | Apr 1999 | A |
5899963 | Hutchings | May 1999 | A |
5941828 | Archibald et al. | Aug 1999 | A |
5947868 | Dugan | Sep 1999 | A |
5954644 | Dettling et al. | Sep 1999 | A |
5955667 | Fyfe | Sep 1999 | A |
5976083 | Richardson et al. | Nov 1999 | A |
6018705 | Gaudet et al. | Jan 2000 | A |
6076015 | Hartley et al. | Jun 2000 | A |
6077193 | Buhler et al. | Jun 2000 | A |
6078874 | Piety et al. | Jun 2000 | A |
6085248 | Sambamurthy et al. | Jul 2000 | A |
6099478 | Aoshima et al. | Aug 2000 | A |
6129686 | Friedman | Oct 2000 | A |
6131076 | Stephan et al. | Oct 2000 | A |
6145389 | Ebeling et al. | Nov 2000 | A |
6183425 | Whalen et al. | Feb 2001 | B1 |
6213872 | Harada et al. | Apr 2001 | B1 |
6241684 | Amano et al. | Jun 2001 | B1 |
6287262 | Amano et al. | Sep 2001 | B1 |
6289230 | Chaiken et al. | Sep 2001 | B1 |
6301964 | Fyfe et al. | Oct 2001 | B1 |
6302789 | Harada et al. | Oct 2001 | B2 |
6305221 | Hutchings | Oct 2001 | B1 |
6307576 | Rosenfeld | Oct 2001 | B1 |
6309360 | Mault | Oct 2001 | B1 |
6360113 | Dettling | Mar 2002 | B1 |
6402690 | Rhee et al. | Jun 2002 | B1 |
6418394 | Puolakanaho et al. | Jul 2002 | B1 |
6469639 | Tanenhaus et al. | Oct 2002 | B2 |
6478736 | Mault | Nov 2002 | B1 |
6513381 | Fyfe et al. | Feb 2003 | B2 |
6513532 | Mault et al. | Feb 2003 | B2 |
6527711 | Stivoric et al. | Mar 2003 | B1 |
6529827 | Beason et al. | Mar 2003 | B1 |
6558335 | Thede | May 2003 | B1 |
6561951 | Cannon et al. | May 2003 | B2 |
6571200 | Mault | May 2003 | B1 |
6583369 | Montagnino et al. | Jun 2003 | B2 |
6585622 | Shum et al. | Jul 2003 | B1 |
6607493 | Song | Aug 2003 | B2 |
6620078 | Pfeffer | Sep 2003 | B2 |
6678629 | Tsuji | Jan 2004 | B2 |
6699188 | Wessel | Mar 2004 | B2 |
6720860 | Narayanaswami | Apr 2004 | B1 |
6731967 | Turcott | May 2004 | B1 |
6761064 | Tsuji | Jul 2004 | B2 |
6790178 | Mault et al. | Sep 2004 | B1 |
6808473 | Hisano et al. | Oct 2004 | B2 |
6811516 | Dugan | Nov 2004 | B1 |
6813582 | Levi et al. | Nov 2004 | B2 |
6813931 | Yadav et al. | Nov 2004 | B2 |
6856938 | Kurtz | Feb 2005 | B2 |
6862575 | Anttila et al. | Mar 2005 | B1 |
6882955 | Ohlenbusch et al. | Apr 2005 | B1 |
6959259 | Vock et al. | Oct 2005 | B2 |
6997882 | Parker et al. | Feb 2006 | B1 |
7020508 | Stivoric et al. | Mar 2006 | B2 |
7041032 | Calvano | May 2006 | B1 |
7062225 | White | Jun 2006 | B2 |
7099237 | Lall | Aug 2006 | B2 |
7133690 | Ranta-Aho et al. | Nov 2006 | B2 |
7153262 | Stivoric et al. | Dec 2006 | B2 |
7162368 | Levi et al. | Jan 2007 | B2 |
7171331 | Vock et al. | Jan 2007 | B2 |
7200517 | Darley et al. | Apr 2007 | B2 |
7246033 | Kudo | Jul 2007 | B1 |
7252639 | Kimura et al. | Aug 2007 | B2 |
7261690 | Teller et al. | Aug 2007 | B2 |
7272982 | Neuhauser et al. | Sep 2007 | B2 |
7285090 | Stivoric et al. | Oct 2007 | B2 |
7334472 | Seo et al. | Feb 2008 | B2 |
7373820 | James | May 2008 | B1 |
7443292 | Jensen et al. | Oct 2008 | B2 |
7457724 | Vock et al. | Nov 2008 | B2 |
7467060 | Kulach et al. | Dec 2008 | B2 |
7502643 | Farringdon et al. | Mar 2009 | B2 |
7505865 | Ohkubo et al. | Mar 2009 | B2 |
7539532 | Tran | May 2009 | B2 |
7558622 | Tran | Jul 2009 | B2 |
7559877 | Parks et al. | Jul 2009 | B2 |
7579946 | Case, Jr. | Aug 2009 | B2 |
7608050 | Shugg | Oct 2009 | B2 |
7653508 | Kahn et al. | Jan 2010 | B1 |
7690556 | Kahn et al. | Apr 2010 | B1 |
7713173 | Shin et al. | May 2010 | B2 |
7720306 | Gardiner et al. | May 2010 | B2 |
7762952 | Lee et al. | Jul 2010 | B2 |
7771320 | Riley et al. | Aug 2010 | B2 |
7774156 | Niva et al. | Aug 2010 | B2 |
7789802 | Lee et al. | Sep 2010 | B2 |
7881902 | Kahn et al. | Feb 2011 | B1 |
7909768 | Turcott | Mar 2011 | B1 |
7927253 | Vincent et al. | Apr 2011 | B2 |
7942824 | Kayyali et al. | May 2011 | B1 |
7953549 | Graham | May 2011 | B2 |
7983876 | Vock et al. | Jul 2011 | B2 |
7993276 | Nazarian et al. | Aug 2011 | B2 |
8005922 | Boudreau | Aug 2011 | B2 |
8028443 | Case, Jr. | Oct 2011 | B2 |
8040758 | Dickinson | Oct 2011 | B1 |
8055469 | Kulach et al. | Nov 2011 | B2 |
8073707 | Teller et al. | Dec 2011 | B2 |
8099318 | Moukas et al. | Jan 2012 | B2 |
8109858 | Redmann | Feb 2012 | B2 |
8132037 | Fehr et al. | Mar 2012 | B2 |
8140143 | Picard et al. | Mar 2012 | B2 |
8152745 | Smith et al. | Apr 2012 | B2 |
8157731 | Teller et al. | Apr 2012 | B2 |
8172761 | Rulkov et al. | May 2012 | B1 |
8177260 | Tropper et al. | May 2012 | B2 |
8180591 | Yuen et al. | May 2012 | B2 |
8180592 | Yuen et al. | May 2012 | B2 |
8199126 | Taubman | Jun 2012 | B1 |
8211503 | Tsao et al. | Jul 2012 | B2 |
8270297 | Akasaka et al. | Sep 2012 | B2 |
8311769 | Yuen et al. | Nov 2012 | B2 |
8311770 | Yuen et al. | Nov 2012 | B2 |
8321004 | Moon et al. | Nov 2012 | B2 |
8346328 | Mannheimer et al. | Jan 2013 | B2 |
8364250 | Moon et al. | Jan 2013 | B2 |
8386008 | Yuen et al. | Feb 2013 | B2 |
8386042 | Yudovsky et al. | Feb 2013 | B2 |
8398546 | Pacione et al. | Mar 2013 | B2 |
8437824 | Moon et al. | May 2013 | B2 |
8437980 | Yuen et al. | May 2013 | B2 |
8444578 | Bourget et al. | May 2013 | B2 |
8446275 | Utter, II | May 2013 | B2 |
8462591 | Marhaben | Jun 2013 | B1 |
8463576 | Yuen et al. | Jun 2013 | B2 |
8463577 | Yuen et al. | Jun 2013 | B2 |
8475367 | Yuen et al. | Jul 2013 | B1 |
8487771 | Hsieh et al. | Jul 2013 | B2 |
8527038 | Moon et al. | Sep 2013 | B2 |
8533269 | Brown | Sep 2013 | B2 |
8533620 | Hoffman et al. | Sep 2013 | B2 |
8543185 | Yuen et al. | Sep 2013 | B2 |
8543351 | Yuen et al. | Sep 2013 | B2 |
8548770 | Yuen et al. | Oct 2013 | B2 |
8554297 | Moon et al. | Oct 2013 | B2 |
8562489 | Burton et al. | Oct 2013 | B2 |
8579827 | Rulkov et al. | Nov 2013 | B1 |
8583402 | Yuen et al. | Nov 2013 | B2 |
8597093 | Engelberg et al. | Dec 2013 | B2 |
8634796 | Johnson | Jan 2014 | B2 |
8641612 | Teller et al. | Feb 2014 | B2 |
8670953 | Yuen et al. | Mar 2014 | B2 |
8684900 | Tran | Apr 2014 | B2 |
8690578 | Nusbaum et al. | Apr 2014 | B1 |
8738118 | Moon et al. | May 2014 | B2 |
8738321 | Yuen et al. | May 2014 | B2 |
8738323 | Yuen et al. | May 2014 | B2 |
8742325 | Droz et al. | Jun 2014 | B1 |
8744803 | Park et al. | Jun 2014 | B2 |
8762101 | Yuen et al. | Jun 2014 | B2 |
8764651 | Tran | Jul 2014 | B2 |
8792981 | Yudovsky et al. | Jul 2014 | B2 |
8827906 | Yuen et al. | Sep 2014 | B2 |
8847988 | Geisner et al. | Sep 2014 | B2 |
8868377 | Yuen et al. | Oct 2014 | B2 |
8909543 | Tropper et al. | Dec 2014 | B2 |
8920332 | Hong et al. | Dec 2014 | B2 |
8936552 | Kateraas et al. | Jan 2015 | B2 |
8938279 | Heaton, II et al. | Jan 2015 | B1 |
8945017 | Venkatraman et al. | Feb 2015 | B2 |
8948832 | Hong et al. | Feb 2015 | B2 |
8949070 | Kahn et al. | Feb 2015 | B1 |
8954135 | Yuen et al. | Feb 2015 | B2 |
8954290 | Yuen et al. | Feb 2015 | B2 |
8956303 | Hong et al. | Feb 2015 | B2 |
8961413 | Teller et al. | Feb 2015 | B2 |
8968195 | Tran | Mar 2015 | B2 |
8998815 | Venkatraman et al. | Apr 2015 | B2 |
9005129 | Venkatraman et al. | Apr 2015 | B2 |
9014790 | Richards et al. | Apr 2015 | B2 |
9031812 | Roberts et al. | May 2015 | B2 |
9039614 | Yuen et al. | May 2015 | B2 |
9042971 | Brumback et al. | May 2015 | B2 |
9044149 | Richards et al. | Jun 2015 | B2 |
9044150 | Brumback et al. | Jun 2015 | B2 |
9047648 | Lekutai et al. | Jun 2015 | B1 |
9049998 | Brumback et al. | Jun 2015 | B2 |
9089760 | Tropper et al. | Jul 2015 | B2 |
9113794 | Hong et al. | Aug 2015 | B2 |
9113795 | Hong et al. | Aug 2015 | B2 |
9226663 | Fei | Jan 2016 | B2 |
9237855 | Hong et al. | Jan 2016 | B2 |
9241646 | Addison et al. | Jan 2016 | B2 |
9282902 | Richards et al. | Mar 2016 | B2 |
9307917 | Hong et al. | Apr 2016 | B2 |
9314166 | Brady et al. | Apr 2016 | B1 |
9314197 | Eisen et al. | Apr 2016 | B2 |
9392946 | Sarantos et al. | Jul 2016 | B1 |
9402552 | Richards et al. | Aug 2016 | B2 |
9456787 | Venkatraman et al. | Oct 2016 | B2 |
9596999 | Moon | Mar 2017 | B2 |
9662053 | Richards et al. | May 2017 | B2 |
9775548 | Sarantos et al. | Oct 2017 | B2 |
10178973 | Venkatraman et al. | Jan 2019 | B2 |
10216893 | Hong et al. | Feb 2019 | B2 |
10216894 | Hong et al. | Feb 2019 | B2 |
10381109 | Hong et al. | Aug 2019 | B2 |
10433739 | Weekly et al. | Oct 2019 | B2 |
10512407 | Richards et al. | Dec 2019 | B2 |
10568525 | Wu et al. | Feb 2020 | B1 |
20010044588 | Mault | Nov 2001 | A1 |
20010049470 | Mault et al. | Dec 2001 | A1 |
20010055242 | Deshmuhk et al. | Dec 2001 | A1 |
20020013717 | Ando et al. | Jan 2002 | A1 |
20020019585 | Dickinson | Feb 2002 | A1 |
20020077219 | Cohen et al. | Jun 2002 | A1 |
20020077536 | Diab et al. | Jun 2002 | A1 |
20020082144 | Pfeffer | Jun 2002 | A1 |
20020087264 | Hills et al. | Jul 2002 | A1 |
20020091329 | Heikkila et al. | Jul 2002 | A1 |
20020109600 | Mault et al. | Aug 2002 | A1 |
20020139936 | Dumas | Oct 2002 | A1 |
20020178060 | Sheehan | Nov 2002 | A1 |
20020198776 | Nara et al. | Dec 2002 | A1 |
20030018523 | Rappaport et al. | Jan 2003 | A1 |
20030050537 | Wessel | Mar 2003 | A1 |
20030065561 | Brown et al. | Apr 2003 | A1 |
20030107487 | Korman et al. | Jun 2003 | A1 |
20030128867 | Bennett | Jul 2003 | A1 |
20030131059 | Brown et al. | Jul 2003 | A1 |
20030163710 | Ortiz et al. | Aug 2003 | A1 |
20030171189 | Kaufman | Sep 2003 | A1 |
20030229276 | Sarussi et al. | Dec 2003 | A1 |
20040054497 | Kurtz | Mar 2004 | A1 |
20040061324 | Howard | Apr 2004 | A1 |
20040117963 | Schneider | Jun 2004 | A1 |
20040152957 | Stivoric et al. | Aug 2004 | A1 |
20040171969 | Socci et al. | Sep 2004 | A1 |
20040190085 | Silverbrook et al. | Sep 2004 | A1 |
20040236227 | Gueissaz | Nov 2004 | A1 |
20040239497 | Schwartzman et al. | Dec 2004 | A1 |
20040249299 | Cobb | Dec 2004 | A1 |
20040257557 | Block | Dec 2004 | A1 |
20050020927 | Blondeau et al. | Jan 2005 | A1 |
20050037844 | Shum et al. | Feb 2005 | A1 |
20050038679 | Short | Feb 2005 | A1 |
20050054938 | Wehman et al. | Mar 2005 | A1 |
20050054940 | Almen | Mar 2005 | A1 |
20050102172 | Sirmans, Jr. | May 2005 | A1 |
20050107723 | Wehman et al. | May 2005 | A1 |
20050163056 | Ranta-Aho et al. | Jul 2005 | A1 |
20050171410 | Hjelt et al. | Aug 2005 | A1 |
20050186965 | Pagonis et al. | Aug 2005 | A1 |
20050187481 | Hatib et al. | Aug 2005 | A1 |
20050195830 | Chitrapu et al. | Sep 2005 | A1 |
20050228244 | Banet | Oct 2005 | A1 |
20050228692 | Hodgdon | Oct 2005 | A1 |
20050234742 | Hodgdon | Oct 2005 | A1 |
20050245793 | Hilton et al. | Nov 2005 | A1 |
20050248718 | Howell et al. | Nov 2005 | A1 |
20050253047 | Maegawa et al. | Nov 2005 | A1 |
20050272564 | Pyles et al. | Dec 2005 | A1 |
20060004265 | Pulkkinen et al. | Jan 2006 | A1 |
20060020174 | Matsumura | Jan 2006 | A1 |
20060020177 | Seo et al. | Jan 2006 | A1 |
20060025282 | Redmann | Feb 2006 | A1 |
20060039348 | Racz et al. | Feb 2006 | A1 |
20060047208 | Yoon | Mar 2006 | A1 |
20060047447 | Brady et al. | Mar 2006 | A1 |
20060052727 | Palestrant | Mar 2006 | A1 |
20060064276 | Ren et al. | Mar 2006 | A1 |
20060069619 | Walker et al. | Mar 2006 | A1 |
20060089542 | Sands | Apr 2006 | A1 |
20060111944 | Sirmans, Jr. | May 2006 | A1 |
20060129436 | Short | Jun 2006 | A1 |
20060143645 | Vock et al. | Jun 2006 | A1 |
20060166718 | Seshadri et al. | Jul 2006 | A1 |
20060189872 | Arnold | Aug 2006 | A1 |
20060195020 | Martin et al. | Aug 2006 | A1 |
20060217231 | Parks et al. | Sep 2006 | A1 |
20060247952 | Muraca | Nov 2006 | A1 |
20060277474 | Robarts et al. | Dec 2006 | A1 |
20060282021 | DeVaul et al. | Dec 2006 | A1 |
20060287883 | Turgiss et al. | Dec 2006 | A1 |
20060288117 | Raveendran | Dec 2006 | A1 |
20070011028 | Sweeney | Jan 2007 | A1 |
20070049384 | Kinq et al. | Mar 2007 | A1 |
20070050715 | Behar | Mar 2007 | A1 |
20070051369 | Choi et al. | Mar 2007 | A1 |
20070061593 | Celikkan et al. | Mar 2007 | A1 |
20070071643 | Hall et al. | Mar 2007 | A1 |
20070072156 | Kaufman et al. | Mar 2007 | A1 |
20070083095 | Rippo et al. | Apr 2007 | A1 |
20070083602 | Heggenhougen et al. | Apr 2007 | A1 |
20070123391 | Shin et al. | May 2007 | A1 |
20070135264 | Rosenberg | Jun 2007 | A1 |
20070136093 | Rankin et al. | Jun 2007 | A1 |
20070146116 | Kimbrell | Jun 2007 | A1 |
20070155277 | Amitai et al. | Jul 2007 | A1 |
20070159926 | Prstojevich et al. | Jul 2007 | A1 |
20070179356 | Wessel | Aug 2007 | A1 |
20070194066 | Ishihara et al. | Aug 2007 | A1 |
20070197920 | Adams | Aug 2007 | A1 |
20070208544 | Kulach et al. | Sep 2007 | A1 |
20070213020 | Novac | Sep 2007 | A1 |
20070219059 | Schwartz et al. | Sep 2007 | A1 |
20070265533 | Tran | Nov 2007 | A1 |
20070276271 | Chan | Nov 2007 | A1 |
20070288265 | Quinian et al. | Dec 2007 | A1 |
20080001735 | Tran | Jan 2008 | A1 |
20080014947 | Carnall | Jan 2008 | A1 |
20080022089 | Leedom | Jan 2008 | A1 |
20080032864 | Hakki | Feb 2008 | A1 |
20080039729 | Cho et al. | Feb 2008 | A1 |
20080054072 | Katragadda | Mar 2008 | A1 |
20080084823 | Akasaka et al. | Apr 2008 | A1 |
20080093838 | Tropper et al. | Apr 2008 | A1 |
20080097221 | Florian | Apr 2008 | A1 |
20080097550 | Dicks et al. | Apr 2008 | A1 |
20080114829 | Button et al. | May 2008 | A1 |
20080125288 | Case | May 2008 | A1 |
20080129457 | Ritter et al. | Jun 2008 | A1 |
20080134102 | Movold et al. | Jun 2008 | A1 |
20080140163 | Keacher et al. | Jun 2008 | A1 |
20080140338 | No et al. | Jun 2008 | A1 |
20080146892 | LeBoeuf et al. | Jun 2008 | A1 |
20080155077 | James | Jun 2008 | A1 |
20080176655 | James et al. | Jul 2008 | A1 |
20080214360 | Stirling et al. | Sep 2008 | A1 |
20080214903 | Orbach | Sep 2008 | A1 |
20080249836 | Angell et al. | Oct 2008 | A1 |
20080275309 | Stivoric et al. | Nov 2008 | A1 |
20080287751 | Stivoric et al. | Nov 2008 | A1 |
20090012433 | Femstrom et al. | Jan 2009 | A1 |
20090018797 | Kasama et al. | Jan 2009 | A1 |
20090043531 | Kahn et al. | Feb 2009 | A1 |
20090047645 | Dibenedetto et al. | Feb 2009 | A1 |
20090048044 | Oleson et al. | Feb 2009 | A1 |
20090054737 | Magar et al. | Feb 2009 | A1 |
20090054751 | Babashan et al. | Feb 2009 | A1 |
20090063193 | Barton et al. | Mar 2009 | A1 |
20090063293 | Mirrashidi et al. | Mar 2009 | A1 |
20090093341 | James et al. | Apr 2009 | A1 |
20090098821 | Shinya | Apr 2009 | A1 |
20090132197 | Rubin et al. | May 2009 | A1 |
20090143655 | Shani | Jun 2009 | A1 |
20090144456 | Gelf et al. | Jun 2009 | A1 |
20090144639 | Nims et al. | Jun 2009 | A1 |
20090150178 | Sutton et al. | Jun 2009 | A1 |
20090156172 | Chan | Jun 2009 | A1 |
20090163783 | Mannheimer et al. | Jun 2009 | A1 |
20090171788 | Tropper et al. | Jul 2009 | A1 |
20090195350 | Tsern et al. | Aug 2009 | A1 |
20090216499 | Tobola et al. | Aug 2009 | A1 |
20090262088 | Moll-Carrillo et al. | Oct 2009 | A1 |
20090264713 | Van Loenen et al. | Oct 2009 | A1 |
20090271147 | Sugai | Oct 2009 | A1 |
20090287921 | Zhu et al. | Nov 2009 | A1 |
20090292332 | Li et al. | Nov 2009 | A1 |
20090307517 | Fehr et al. | Dec 2009 | A1 |
20090309742 | Alexander et al. | Dec 2009 | A1 |
20090318779 | Tran | Dec 2009 | A1 |
20100023348 | Hardee et al. | Jan 2010 | A1 |
20100026995 | Merritt et al. | Feb 2010 | A1 |
20100059561 | Ellis et al. | Mar 2010 | A1 |
20100063365 | Pisani et al. | Mar 2010 | A1 |
20100069203 | Kawaguchi et al. | Mar 2010 | A1 |
20100079291 | Kroll et al. | Apr 2010 | A1 |
20100106044 | Linderman | Apr 2010 | A1 |
20100113948 | Yang et al. | May 2010 | A1 |
20100125729 | Baentsch et al. | May 2010 | A1 |
20100130873 | Yuen et al. | May 2010 | A1 |
20100152600 | Droitcour et al. | Jun 2010 | A1 |
20100158494 | King | Jun 2010 | A1 |
20100159709 | Kotani et al. | Jun 2010 | A1 |
20100167783 | Alameh et al. | Jul 2010 | A1 |
20100179411 | Holmstrom et al. | Jul 2010 | A1 |
20100185064 | Bandic et al. | Jul 2010 | A1 |
20100204550 | Heneghan et al. | Aug 2010 | A1 |
20100205541 | Rapaport et al. | Aug 2010 | A1 |
20100217099 | LeBoeuf et al. | Aug 2010 | A1 |
20100222179 | Temple et al. | Sep 2010 | A1 |
20100240972 | Neal | Sep 2010 | A1 |
20100249633 | Droitcour et al. | Sep 2010 | A1 |
20100261987 | Kamath et al. | Oct 2010 | A1 |
20100274100 | Behar et al. | Oct 2010 | A1 |
20100292050 | DiBenedetto | Nov 2010 | A1 |
20100292568 | Droitcour et al. | Nov 2010 | A1 |
20100292600 | DiBenedetto et al. | Nov 2010 | A1 |
20100295684 | Hsieh et al. | Nov 2010 | A1 |
20100298650 | Moon et al. | Nov 2010 | A1 |
20100298651 | Moon et al. | Nov 2010 | A1 |
20100298656 | Mccombie et al. | Nov 2010 | A1 |
20100298661 | McCombie et al. | Nov 2010 | A1 |
20100311544 | Robinette et al. | Dec 2010 | A1 |
20100324384 | Moon et al. | Dec 2010 | A1 |
20100331145 | Lakovic et al. | Dec 2010 | A1 |
20100331657 | Mensinger et al. | Dec 2010 | A1 |
20110003665 | Burton et al. | Jan 2011 | A1 |
20110009051 | Khedouri et al. | Jan 2011 | A1 |
20110009727 | Mensinger et al. | Jan 2011 | A1 |
20110022349 | Stirling et al. | Jan 2011 | A1 |
20110032105 | Hoffman et al. | Feb 2011 | A1 |
20110066009 | Moon et al. | Mar 2011 | A1 |
20110066010 | Moon et al. | Mar 2011 | A1 |
20110066044 | Moon et al. | Mar 2011 | A1 |
20110080349 | Holbein et al. | Apr 2011 | A1 |
20110087076 | Brynelsen et al. | Apr 2011 | A1 |
20110106449 | Chowdhary et al. | May 2011 | A1 |
20110112442 | Meger et al. | May 2011 | A1 |
20110118621 | Chu | May 2011 | A1 |
20110131005 | Ueshima et al. | Jun 2011 | A1 |
20110153773 | Vandwalle | Jun 2011 | A1 |
20110167262 | Ross et al. | Jul 2011 | A1 |
20110184248 | Furuta | Jul 2011 | A1 |
20110193704 | Harper et al. | Aug 2011 | A1 |
20110197157 | Hoffman et al. | Aug 2011 | A1 |
20110214030 | Greenberg et al. | Sep 2011 | A1 |
20110224508 | Moon | Sep 2011 | A1 |
20110230729 | Shirasaki et al. | Sep 2011 | A1 |
20110237911 | Lamego et al. | Sep 2011 | A1 |
20110237912 | Couronne et al. | Sep 2011 | A1 |
20110263950 | Larson et al. | Oct 2011 | A1 |
20110276304 | Yin et al. | Nov 2011 | A1 |
20120035487 | Werner et al. | Feb 2012 | A1 |
20120072165 | Jallon | Mar 2012 | A1 |
20120083705 | Yuen et al. | Apr 2012 | A1 |
20120083714 | Yuen et al. | Apr 2012 | A1 |
20120083715 | Yuen et al. | Apr 2012 | A1 |
20120083716 | Yuen et al. | Apr 2012 | A1 |
20120084053 | Yuen et al. | Apr 2012 | A1 |
20120084054 | Yuen et al. | Apr 2012 | A1 |
20120092157 | Tran | Apr 2012 | A1 |
20120094649 | Porrati et al. | Apr 2012 | A1 |
20120116684 | Ingrassia, Jr. et al. | May 2012 | A1 |
20120119911 | Jeon et al. | May 2012 | A1 |
20120123232 | Najarian et al. | May 2012 | A1 |
20120140233 | Rockwell et al. | Jun 2012 | A1 |
20120143067 | Watson et al. | Jun 2012 | A1 |
20120150052 | Buccheim et al. | Jun 2012 | A1 |
20120150074 | Yanev et al. | Jun 2012 | A1 |
20120165684 | Sholder | Jun 2012 | A1 |
20120166257 | Shiraqarni et al. | Jun 2012 | A1 |
20120172733 | Park | Jul 2012 | A1 |
20120179011 | Moon et al. | Jul 2012 | A1 |
20120179278 | Riley et al. | Jul 2012 | A1 |
20120183939 | Aragones et al. | Jul 2012 | A1 |
20120215328 | Schmelzer | Aug 2012 | A1 |
20120226471 | Yuen et al. | Sep 2012 | A1 |
20120226472 | Yuen et al. | Sep 2012 | A1 |
20120227737 | Mastrototaro et al. | Sep 2012 | A1 |
20120229270 | Morley et al. | Sep 2012 | A1 |
20120232432 | Kahn et al. | Sep 2012 | A1 |
20120245439 | Andre et al. | Sep 2012 | A1 |
20120253486 | Niemimaki | Oct 2012 | A1 |
20120255875 | Vicente et al. | Oct 2012 | A1 |
20120265480 | Oshima | Oct 2012 | A1 |
20120271121 | Della Torre et al. | Oct 2012 | A1 |
20120271180 | Ren et al. | Oct 2012 | A1 |
20120274508 | Brown et al. | Nov 2012 | A1 |
20120283855 | Hoffman et al. | Nov 2012 | A1 |
20120296400 | Bierman et al. | Nov 2012 | A1 |
20120297229 | Desai et al. | Nov 2012 | A1 |
20120316456 | Rahman et al. | Dec 2012 | A1 |
20120316471 | Rahman et al. | Dec 2012 | A1 |
20120324226 | Bichsel et al. | Dec 2012 | A1 |
20120330109 | Tran | Dec 2012 | A1 |
20130006718 | Nielsen et al. | Jan 2013 | A1 |
20130009779 | Wittling et al. | Jan 2013 | A1 |
20130041590 | Burich et al. | Feb 2013 | A1 |
20130053661 | Alberth et al. | Feb 2013 | A1 |
20130072169 | Ross et al. | Mar 2013 | A1 |
20130073254 | Yuen et al. | Mar 2013 | A1 |
20130073255 | Yuen et al. | Mar 2013 | A1 |
20130077823 | Mestha et al. | Mar 2013 | A1 |
20130077826 | Cowperthwaite et al. | Mar 2013 | A1 |
20130079607 | Gareau et al. | Mar 2013 | A1 |
20130080113 | Yuen et al. | Mar 2013 | A1 |
20130094600 | Beziat et al. | Apr 2013 | A1 |
20130095459 | Tran | Apr 2013 | A1 |
20130096843 | Yuen et al. | Apr 2013 | A1 |
20130102251 | Linde et al. | Apr 2013 | A1 |
20130106684 | Weast et al. | May 2013 | A1 |
20130132501 | Vandwalle et al. | May 2013 | A1 |
20130151196 | Yuen et al. | Jun 2013 | A1 |
20130158369 | Yuen et al. | Jun 2013 | A1 |
20130163390 | Gossweiler, III et al. | Jun 2013 | A1 |
20130166048 | Werner et al. | Jun 2013 | A1 |
20130173171 | Drysdale et al. | Jul 2013 | A1 |
20130190903 | Balakrishnan et al. | Jul 2013 | A1 |
20130191034 | Weast et al. | Jul 2013 | A1 |
20130203475 | Kil et al. | Aug 2013 | A1 |
20130209972 | Carter et al. | Aug 2013 | A1 |
20130211265 | Bedingham et al. | Aug 2013 | A1 |
20130218053 | Kaiser et al. | Aug 2013 | A1 |
20130225117 | Giacoletto et al. | Aug 2013 | A1 |
20130228063 | Turner | Sep 2013 | A1 |
20130231574 | Tran | Sep 2013 | A1 |
20130238287 | Hoffman et al. | Sep 2013 | A1 |
20130245436 | Tupin, Jr. et al. | Sep 2013 | A1 |
20130261415 | Ashe et al. | Oct 2013 | A1 |
20130261475 | Mochizuki | Oct 2013 | A1 |
20130267249 | Rosenberg | Oct 2013 | A1 |
20130268199 | Nielsen et al. | Oct 2013 | A1 |
20130268236 | Yuen et al. | Oct 2013 | A1 |
20130268687 | Schrecker | Oct 2013 | A1 |
20130274904 | Coza et al. | Oct 2013 | A1 |
20130280682 | Levine et al. | Oct 2013 | A1 |
20130281110 | Zelinka | Oct 2013 | A1 |
20130289366 | Chua et al. | Oct 2013 | A1 |
20130296666 | Kumar et al. | Nov 2013 | A1 |
20130296672 | O'Neil et al. | Nov 2013 | A1 |
20130296673 | Thaveeprungsriporn et al. | Nov 2013 | A1 |
20130296714 | Kassim | Nov 2013 | A1 |
20130310896 | Mass | Nov 2013 | A1 |
20130325396 | Yuen et al. | Dec 2013 | A1 |
20130331058 | Harvey | Dec 2013 | A1 |
20130337974 | Yanev et al. | Dec 2013 | A1 |
20140035761 | Burton et al. | Feb 2014 | A1 |
20140039284 | Niwayama et al. | Feb 2014 | A1 |
20140039804 | Park et al. | Feb 2014 | A1 |
20140039840 | Yuen et al. | Feb 2014 | A1 |
20140039841 | Yuen et al. | Feb 2014 | A1 |
20140052280 | Yuen et al. | Feb 2014 | A1 |
20140067278 | Yuen et al. | Mar 2014 | A1 |
20140073486 | Ahmed | Mar 2014 | A1 |
20140074431 | Modi | Mar 2014 | A1 |
20140077673 | Garg et al. | Mar 2014 | A1 |
20140088385 | Moon et al. | Mar 2014 | A1 |
20140094941 | Ellis et al. | Apr 2014 | A1 |
20140099614 | Hu et al. | Apr 2014 | A1 |
20140107493 | Yuen et al. | Apr 2014 | A1 |
20140125618 | Panther et al. | May 2014 | A1 |
20140135612 | Yuen et al. | May 2014 | A1 |
20140135631 | Brumback et al. | May 2014 | A1 |
20140142403 | Brumback et al. | May 2014 | A1 |
20140164611 | Molettiere et al. | Jun 2014 | A1 |
20140180022 | Stivoric et al. | Jun 2014 | A1 |
20140207264 | Quy | Jul 2014 | A1 |
20140213858 | Presura et al. | Jul 2014 | A1 |
20140228649 | Rayner | Aug 2014 | A1 |
20140241626 | Sull et al. | Aug 2014 | A1 |
20140275821 | Beckman | Sep 2014 | A1 |
20140275852 | Hong et al. | Sep 2014 | A1 |
20140275854 | Venkatraman et al. | Sep 2014 | A1 |
20140275885 | Isaacson et al. | Sep 2014 | A1 |
20140276119 | Venkatraman et al. | Sep 2014 | A1 |
20140278139 | Hong et al. | Sep 2014 | A1 |
20140278229 | Hong et al. | Sep 2014 | A1 |
20140288390 | Hong et al. | Sep 2014 | A1 |
20140288391 | Hong et al. | Sep 2014 | A1 |
20140288392 | Hong et al. | Sep 2014 | A1 |
20140288435 | Richards et al. | Sep 2014 | A1 |
20140288436 | Venkatraman et al. | Sep 2014 | A1 |
20140288438 | Venkatraman et al. | Sep 2014 | A1 |
20140303523 | Hong et al. | Oct 2014 | A1 |
20140316305 | Venkatraman et al. | Oct 2014 | A1 |
20140323827 | Ahmed et al. | Oct 2014 | A1 |
20140323828 | Ahmed et al. | Oct 2014 | A1 |
20140343372 | Ahmed et al. | Nov 2014 | A1 |
20140350356 | Ahmed et al. | Nov 2014 | A1 |
20140378786 | Hong et al. | Dec 2014 | A1 |
20140378787 | Brumback et al. | Dec 2014 | A1 |
20140378844 | Fei | Dec 2014 | A1 |
20140378872 | Hong et al. | Dec 2014 | A1 |
20150025393 | Hong et al. | Jan 2015 | A1 |
20150025394 | Hong et al. | Jan 2015 | A1 |
20150026647 | Park et al. | Jan 2015 | A1 |
20150173628 | Yuen et al. | Jun 2015 | A1 |
20150173631 | Richards et al. | Jun 2015 | A1 |
20150196256 | Venkatraman et al. | Jul 2015 | A1 |
20150201853 | Hong et al. | Jul 2015 | A1 |
20150201854 | Hong et al. | Jul 2015 | A1 |
20150223708 | Richards et al. | Aug 2015 | A1 |
20150230743 | Silveira et al. | Aug 2015 | A1 |
20150230761 | Brumback et al. | Aug 2015 | A1 |
20150282713 | Fei | Oct 2015 | A1 |
20150351646 | Cervini | Dec 2015 | A1 |
20150366469 | Harris et al. | Dec 2015 | A1 |
20150366504 | Connor et al. | Dec 2015 | A1 |
20160029968 | Lerner et al. | Feb 2016 | A1 |
20160034634 | Hong et al. | Feb 2016 | A9 |
20160058309 | Han | Mar 2016 | A1 |
20160058312 | Han et al. | Mar 2016 | A1 |
20160113585 | Uedaira et al. | Apr 2016 | A1 |
20160183818 | Richards et al. | Jun 2016 | A1 |
20160302706 | Richards et al. | Oct 2016 | A1 |
20160345881 | Sarantos et al. | Dec 2016 | A1 |
20160374567 | Breslow et al. | Dec 2016 | A1 |
20160374569 | Breslow et al. | Dec 2016 | A1 |
20170020659 | Hyde et al. | Jan 2017 | A1 |
20170027523 | Venkatraman et al. | Feb 2017 | A1 |
20170164848 | Nadeau et al. | Jun 2017 | A1 |
20170311825 | Weekly et al. | Nov 2017 | A1 |
20180108802 | Chen | Apr 2018 | A1 |
20180310846 | Lin | Nov 2018 | A1 |
20190082985 | Hong et al. | Mar 2019 | A1 |
20190385708 | Hong et al. | Dec 2019 | A1 |
20200138309 | Weekly et al. | May 2020 | A1 |
Number | Date | Country |
---|---|---|
1623175 | Jun 2005 | CN |
1729933 | Aug 2006 | CN |
101039617 | Sep 2007 | CN |
100362963 | Jan 2008 | CN |
101615098 | Dec 2009 | CN |
101730503 | Jun 2010 | CN |
101742981 | Jun 2010 | CN |
101940476 | Jan 2011 | CN |
102008811 | Apr 2011 | CN |
202069586 | Dec 2011 | CN |
102389313 | Mar 2012 | CN |
102551686 | Jul 2012 | CN |
102750015 | Oct 2012 | CN |
102781310 | Nov 2012 | CN |
103093420 | May 2013 | CN |
104379055 | Feb 2015 | CN |
1 297 784 | Apr 2003 | EP |
1 586 353 | Oct 2005 | EP |
1 721 237 | Aug 2012 | EP |
11-347021 | Dec 1999 | JP |
2010-169410 | May 2010 | JP |
2178588 | Jan 2002 | RU |
WO 02011019 | Feb 2002 | WO |
WO 2006044677 | Apr 2006 | WO |
WO 06055125 | May 2006 | WO |
WO 06090197 | Aug 2006 | WO |
WO 08038141 | Apr 2008 | WO |
WO 09042965 | Apr 2009 | WO |
WO 12061438 | May 2012 | WO |
WO 12170586 | Dec 2012 | WO |
WO 12170924 | Dec 2012 | WO |
WO 12171032 | Dec 2012 | WO |
WO 2014091424 | Jun 2014 | WO |
WO 2014091424 | Jun 2014 | WO |
WO 15127067 | Aug 2015 | WO |
WO 16003269 | Jan 2016 | WO |
WO 2017190051 | Nov 2017 | WO |
Entry |
---|
“Fitbit automatically tracks your fitness and sleep,” Fitbit Inc., published online at www.fitbit.com, downloaded Sep. 10, 2008, 1 page. |
“A Hybrid Discriminative/Generative Approach for Modeling Human Activities”, Lester, et al., Proc. of the Int'l Joint Conf. Artificial intelligence, 2005, pp. 766-772. |
“Activity Classification Using Realistic Data From Wearable Sensors”, Parkka, et al, IEEE Transactions on Information Technology in Biomedicine, vol. 10, No. 1, Jan. 2006, pp. 119-128. |
“Altimeter and Barometer System”, Clifford, et al., Freescale Semiconductor Aplication Note AN1979, Rev. 3, Nov. 2006. |
“An Intelligent Multi-Sensor system for Pedestrian Navigation”, Retscher, Journal of Global Positioning Systems, vol. 5, No. 1, 2006, pp. 110-118. |
“Automatic classification of ambulatory movements and evaluation of energy consumptions utilizing accelerometers and barometer”, Ohtaki, et al, Microsystem Technologies, vol. 11, No. 8-10, Aug. 2005, pp. 1034-1040. |
“Classification of Human Moving Patterns Using Air Pressure and Acceleration”, Sagawa, et al, Proceedings of the 24th Annual Conference of the IEEE Industrial Electronics Society, vol. 2, Aug.-Sep. 1998, pp. 1214-1219. |
“Design of a Wireless Assisted Pedestrian Dead Reckoning System—The NavMote Experience”, Fang, et al, IEEE Transactions on instrumentation and Measurement, vol. 54, No. 6, Dec. 2005, pp. 2342-2358. |
“Direct Measurement of Human Movement by Accelerometry”, Godfrey, et al., Medical Engineering & Physics, vol. 30, 2008, pp. 1364-1386. |
“Drift-free dynamic height sensor using MEMS IMU aided by MEMS pressure sensor”, Tanigawa, et al, Workshop on Positioning, Navigation and Communication, Mar. 2008, pp. 191-196. |
“Evaluation of a New Method of Heading Estimation of Pedestrian Dead Reckoning Using Shoe Mounted Sensors”, Stirling et al., Journal of Navigation, vol. 58, 2005, pp. 31-45. |
“Foot Mounted Inertia System for Pedestrian Naviation”, Godha et al., Measurement Science and Technology, vol. 19, No. 7, May 2008, pp. 1-9. |
“Improvement of Walking Speed Prediction by Accelerometry and Altimetry, Validated by Satellite Positioning”, Perrin, et al, Medical & Biological Engineering & Computing, vol. 38, 2000, pp. 164-168. |
“Indoor Navigation with MEMS Sensors”, Lammel, et al., Proceedings of the Eurosensors XIII conference, vol. 1, No. 1, Sep. 2009, pp. 532-535. |
“Non-restricted measurement of walking distance”, Sagawa, et al, IEEE Int'l Conf. on Systems, Man, and Cybernetics, vol. 3, Oct. 2000, pp. 1847-1852. |
“On Foot Navigation: When GPS alone is not Enough”, Ladetto, et al, Journal of Navigation, vol. 53, No. 2, Sep. 2000, pp. 279-285. |
“SCP 1000-D01/D11 Pressure Sensor as Barometer and Altimeter”, VTI Technologies Application, Jun. 2006, Note 33. |
“Suunto LUMI User Guide”, Jun. and Sep. 1997. |
“Using MS5534 for altimeters and barometers”, Intersema App., Note AN501, Jan. 2006. |
“Validated caloric expenditure estimation using a single body-worn sensor”, Lester, et al, Proc. of the Int'l Conf. on Ubiquitous Computing, 2009, pp. 225-234. |
Deepak et al., Plug-and-Play, Single-Chip Photoplethysmography, 34th Annual International Conference of the IEEE EMBS, San Diego, California USA, Aug. 28-Sep. 1, 2012, 4 pages. Jun. 12, 2014 9:37:30 AM PDT No No. |
Specification of the Bluetooth™ System, Core Package, version 4.1, Dec. 2013, vols. 0 & 1, 282 pages. |
International Search Report dated Aug. 15, 2008, in related application PCT/IB07/03617. |
U.S. Office Action, dated Aug. 4, 2014, issued in U.S. Appl. No. 13/924,784. |
U.S. Notice of Allowance, dated Nov. 19, 2014, issued in U.S. Appl. No. 13/924,784. |
U.S. Office Action, dated Oct. 22, 2014, issued in U.S. Appl. No. 14/290,884. |
U.S. Notice of Allowance, dated Feb. 6, 2015, issued in U.S. Appl. No. 14/290,884. |
U.S. Office Action, dated Jun. 22, 2015, issued in U.S. Appl. No. 14/693,710. |
U.S. Notice of Allowance, dated Jul. 27, 2015, issued in U.S. Appl. No. 14/693,710. |
U.S. Notice of Allowance, dated Apr. 15, 2016, issued in U.S. Appl. No. 14/954,753. |
U.S. Office Action, dated Oct. 26, 2016, issued in U.S. Appl. No. 15/195,911. |
U.S. Notice of Allowance, dated Jan. 23, 2017, issued in U.S. Appl. No. 15/195,911. |
U.S. Notice of Allowance, dated Sep. 23, 2014, issued in U.S. Appl. No. 14/292,669. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Oct. 14, 2014, issued in U.S. Appl. No. 14/292,669. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Dec. 31, 2014, issued in U.S. Appl. No. 14/292,669. |
U.S. Notice of Allowance, dated Oct. 14, 2014, issued in U.S. Appl. No. 14/295,144. |
U.S. Notice of Allowance, dated Dec. 3, 2014, issued in U.S. Appl. No. 14/295,144. |
U.S. Notice of Allowance, dated Sep. 26, 2014, issued in U.S. Appl. No. 14/295,158. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Dec. 31, 2014, issued in U.S. Appl. No. 14/295,158. |
U.S. Office Action, dated Jan. 23, 2015, issued in U.S. Appl. No. 14/507,184. |
U.S. Final Office Action, dated May 11, 2015, issuedin U.S. Appl. No. 14/507,184. |
U.S. Notice of Allowance, dated Aug. 11, 2015, issued in U.S. Appl. No. 14/507,184. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Dec. 18, 2015, issued in U.S. Appl. No. 14/507,184. |
U.S. Office Action, dated Jan. 26, 2015, issued in U.S. Appl. No. 14/295,161. |
U.S. Notice of Allowance, dated Apr. 14, 2015, issued in U.S. Appl. No. 14/295,161. |
U.S. Notice of Allowance, dated Jul. 28, 2015, issued in U.S. Appl. No. 14/295,161. |
U.S. Office Action, dated May 11, 2015, issued in U.S. Appl. No. 14/673,630. |
U.S. Notice of Allowance, dated Nov. 25, 2015, issued in U.S. Appl. No. 14/673,630. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Mar. 21, 2016, issued in U.S. Appl. No. 14/673,630. |
U.S. Office Action, dated Jan. 27, 2015, issued in U.S. Appl. No. 14/507,173. |
U.S. Notice of Allowance, dated Apr. 17, 2015, issued in U.S. Appl. No. 14/507,173. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Jul. 16, 2015, issued in U.S. Appl. No. 14/507,173. |
U.S. Office Action, dated Jun. 8, 2015, issued in U.S. Appl. No. 14/673,634. |
U.S. Final Office Action, dated Nov. 4, 2015, issued in U.S. Appl. No. 14/673,634. |
U.S. Office Action, dated Jul. 13, 2016, issued in U.S. Appl. No. 14/673,634. |
U.S. Office Action, dated Feb. 9, 2017, issued in U.S. Appl. No. 14/673,634. |
U.S. Final Office Action, dated Aug. 9, 2017, issued in U.S. Appl. No. 14/673,634. |
U.S. Office Action, dated Mar. 27, 2018, issued in U.S. Appl. No. 14/673,634. |
U.S. Office Action, dated Aug. 5, 2014, issued in U.S. Appl. No. 14/292,673. |
U.S. Notice of Allowance, dated Dec. 8, 2014, issued in U.S. Appl. No. 14/292,673. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Mar. 5, 2015, issued in U.S. Appl. No. 14/292,673. |
U.S. Office Action, dated Sep. 18, 2014, issued in U.S. Appl. No. 14/295,059. |
U.S. Notice of Allowance, dated Jan. 28, 2015, issued in U.S. Appl. No. 14/295,059. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Mar. 11, 2015, issued in U.S. Appl. No. 14/295,059. |
U.S. Office Action, dated Dec. 24, 2014, issued in U.S. Appl. No. 14/295,076. |
U.S. Final Office Action, dated Apr. 15, 2015, issued in U.S. Appl. No. 14/295,076. |
U.S. Office Action, dated Oct. 22, 2015, issued in U.S. Appl. No. 14/295,076. |
U.S. Notice of Allowance, dated May 24, 2016, issued in U.S. Appl. No. 14/295,076. |
U.S. Office Action, dated Jan. 12, 2018, issued in U.S. Appl. No. 15/246,387. |
U.S. Notice of Allowance, dated Aug. 29, 2018, issued in U.S. Appl. No. 15/246,387. |
U.S. Office Action, dated Jul. 31, 2014, issued in U.S. Appl. No. 14/295,122. |
U.S. Notice of Allowance, dated Nov. 24, 2014, issued in U.S. Appl. No. 14/295,122. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated Jan. 5, 2015, issued in U.S. Appl. No. 14/295,122. |
U.S. Office Action dated Dec. 22, 2016, issued in U.S. Appl. No. 14/599,039. |
U.S. Final Office Action dated Aug. 3, 2017, issued in U.S. Appl. No. 14/599,039. |
U.S. Office Action, dated Mar. 14, 2014, issued in U.S. Appl. No. 14/154,009. |
U.S. Office Action, dated Sep. 29, 2014, issued in U.S. Appl. No. 14/154,009. |
U.S. Notice of Allowance, dated Jan. 21, 2015, issued in U.S. Appl. No. 14/154,009. |
U.S. Office Action, dated Nov. 25, 2014, issued in U.S. Appl. No. 14/154,019. |
U.S. Notice of Allowance, dated Mar. 20, 2015, issued in U.S. Appl. No. 14/154,019. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated May 14, 2015, issued in U.S. Appl. No. 14/154,019. |
U.S. Office Action, dated Jul. 24, 2018, issued in U.S. Appl. No. 14/696,256. |
U.S. Final Office Action, dated Feb. 26, 2019, issued in U.S. Appl. No. 14/696,256. |
U.S. Office Action, dated Feb. 19, 2020, issued in U.S. Appl. No. 14/696,256. |
U.S. Office Action, dated Dec. 10, 2014, issued in U.S. Appl. No. 14/484,104. |
U.S. Notice of Allowance, dated Mar. 19, 2015, issued in U.S. Appl. No. 14/484,104. |
U.S. Notice of Allowance (Corrected Notice of Allowability), dated May 6, 2015, issued in U.S. Appl. No. 14/484,104. |
U.S. Office Action, dated Dec. 4, 2014, issued in U.S. Appl. No. 14/216,743. |
U.S. Final Office Action, dated Apr. 8, 2015, issued in U.S. Appl. No. 14/216,743. |
U.S. Office Action, dated Oct. 2, 2015, issued in U.S. Appl. No. 14/216,743. |
U.S. Final Office Action, dated Feb. 8, 2016, issued in U.S. Appl. No. 14/216,743. |
U.S. Office Action, dated May 16, 2016, issued in U.S. Appl. No. 14/216,743. |
U.S. Office Action, dated Jan. 13, 2017, issued in U.S. Appl. No. 14/216,743. |
U.S. Examiner's Answer to Appeal Brief before the Patent Trial and Appeal Board [in response to the appeal brief filed Sep. 12, 2017 appealing from the Office action dated Jan. 3, 2017], dated Nov. 30, 2017, issued in U.S. Appl. No. 14/216,743. |
U.S. Patent Trial and Appeal Board's Decision on Appeal, dated Oct. 9, 2018, issued in U.S. Appl. No. 14/216,743. |
U.S. Notice of Allowance, dated Dec. 17, 2018, issued in U.S. Appl. No. 14/216,743. |
U.S. Office Action, dated Mar. 12, 2015, issued in U.S. Appl. No. 14/481,020. |
U.S. Final Office Action, dated Jul. 7, 2015, issued in U.S. Appl. No. 14/481,020. |
U.S. Office Action, dated Oct. 27, 2015, issued in U.S. Appl. No. 14/481,020. |
U.S. Final Office Action, dated May 13, 2016, issued in U.S. Appl. No. 14/481,020. |
U.S. Examiner's Answer to Appeal Brief before the Patent Trial and Appeal Board [in response to the appeal brief filed Dec. 9, 2016 appealing from the Office action dated May 13, 2016], dated Jan. 23, 2017, issued in U.S. Appl. No. 14/481,020. |
U.S. Patent Trial and Appeal Board's Decision on Appeal, dated Sep. 14, 2018, issued in U.S. Appl. No. 14/481,020. |
U.S. Notice of Allowance, dated Nov. 29, 2018, issued in U.S. Appl. No. 14/481,020. |
U.S. Office Action, dated Aug. 22, 2014, issued in U.S. Appl. No. 14/250,256. |
U.S. Final Office Action, dated Nov. 21, 2014, issued in U.S. Appl. No. 14/250,256. |
U.S. Office Action, dated Jul. 8, 2015, issued in U.S. Appl. No. 14/250,256. |
U.S. Final Office Action, dated Oct. 23, 2015, issued in U.S. Appl. No. 14/250,256. |
U.S. Office Action, dated Mar. 17, 2016, issued in U.S. Appl. No. 14/250,256. |
U.S. Final Office Action, dated Jun. 29, 2016, issued in U.S. Appl. No. 14/250,256. |
U.S. Office Action, dated Jan. 9, 2017, issued in U.S. Appl. No. 14/250,256. |
U.S. Examiner's Answer to the Appeal Brief before the Patent Trial and Appeal Board [in response to the appeal brief filed Jul. 11, 2017 appealing from the Office action dated Jan. 9, 2017], dated Aug. 24, 2017, issued in U.S. Appl. No. 14/250,256. |
U.S. Patent Trial and Appeal Board's Decision on Appeal, dated Oct. 9, 2018, issued in U.S. Appl. No. 14/250,256. |
U.S. Notice of Allowance, dated Mar. 29, 2019, issued in U.S. Appl. No. 14/250,256. |
U.S. Office Action, dated Oct. 7, 2014, issued in U.S. Appl. No. 14/481,762. |
U.S. Final Office Action, dated Dec. 19, 2014, issued in U.S. Appl. No. 14/481,762. |
U.S. Office Action, dated Jul. 7, 2015, issued in U.S. Appl. No. 14/481,762. |
U.S. Final Office Action, dated Nov. 5, 2015, issued in U.S. Appl. No. 14/481,762. |
U.S. Office Action, dated May 11, 2016, issued in U.S. Appl. No. 14/481,762. |
U.S. Final Office Action, dated Oct. 19, 2016, issued in U.S. Appl. No. 14/481,762. |
U.S. Office Action, dated Apr. 12, 2017, issued in U.S. Appl. No. 14/481,762. |
U.S. Office Action, dated Nov. 19, 2015, issued in U.S. Appl. No. 14/724,750. |
U.S. Notice of Allowance, dated Mar. 8, 2016, issued in U.S. Appl. No. 14/724,750. |
U.S. Office Action dated Sep. 8, 2016, issued in U.S. Appl. No. 15/192,447. |
U.S. Final Office Action dated Feb. 7, 2017, issued in U.S. Appl. No. 15/192,447. |
U.S. Notice of Allowance dated May 24, 2017, issued in U.S. Appl. No. 15/192,447. |
U.S. Office Action dated Mar. 15, 2017, issued in U.S. Appl. No. 15/370,303. |
U.S. Final Office Action dated Aug. 1, 2017, issued in U.S. Appl. No. 15/370,303. |
U.S. Office Action dated Jan. 11, 2018, issued in U.S. Appl. No. 15/370,303. |
U.S. Final Office Action dated Jul. 25, 2018, issued in U.S. Appl. No. 15/370,303. |
U.S. Office Action dated May 24, 2019, issued in U.S. Appl. No. 15/370,303. |
U.S. Office Action, dated Oct. 7, 2014, issued in U.S. Appl. No. 14/292,844. |
U.S. Notice of Allowance, dated Feb. 9, 2015, issued in U.S. Appl. No. 14/292,844. |
U.S. Office Action, dated Jul. 6, 2015, issued in U.S. Appl. No. 14/640,281. |
U.S. Final Office Action, dated Nov. 12, 2015, issued in U.S. Appl. No. 14/640,281. |
U.S. Office Action, dated Oct. 6, 2016, issued in U.S. Appl. No. 14/640,281. |
U.S. Final Office Action, dated May 4, 2017, issued in U.S. Appl. No. 14/640,281. |
U.S. Office Action, dated Jun. 29, 2018, issued in U.S. Appl. No. 14/640,281. |
U.S. Final Office Action, dated Feb. 21, 2019, issued in U.S. Appl. No. 14/640,281. |
U.S. Notice of Allowance, dated Aug. 2, 2019, issued in U.S. Appl. No. 14/640,281. |
U.S. Office Action, dated May 30, 2019, issued in U.S. Appl. No. 15/376,542. |
U.S. Notice of Allowance, dated Sep. 20, 2019, issued in U.S. Appl. No. 15/376,542. |
U.S. Office Action, dated Mar. 11, 2019, issued in U.S. Appl. No. 15/582,240. |
U.S. Notice of Allowance, dated Jun. 14, 2019, issued in U.S. Appl. No. 15/582,240. |
U.S. Office Action, dated Apr. 1, 2020, issued in U.S. Appl. No. 15/948,970. |
Chinese First Office Action dated Sep. 27, 2016 issued in Application No. CN 201410018701.8. |
Chinese Second Office Action dated Jun. 13, 2017 issued in Application No. CN 201410018701.8. |
Chinese First Office Action dated Aug. 7, 2015 issued in Application No. CN 201410243180.6. |
Chinese First Office Action dated Sep. 2, 2016 issued in Application No. CN 201510745382.5. |
Chinese Second Office Action dated Mar. 22, 2017 issued in Application No. CN 201510745382.5. |
Chinese First Office Action dated Mar. 22, 2018 issued in Application No. CN 201610284612.7. |
Chinese Second Office Action dated Nov. 6, 2018 issued in Application No. CN 201610284612.7. |
Chinese First Office Action dated Aug. 3, 2016 issued in Application No. CN 201410243169.X. |
Chinese Second Office Action dated Mar. 27, 2017 issued in Application No. CN 201410243169.X. |
Chinese Third Office Action dated Sep. 28, 2017 issued in Application No. CN 201410243169.X. |
Chinese First Office Action dated Sep. 26, 2016 issued in Application No. CN 201410243178.9. |
Chinese Second Office Action dated Jun. 15, 2017 issued in Application No. CN 201410243178.9. |
Chinese First Office Action dated Mar. 3, 2017 issued in Application No. CN 201610622453.7. |
Chinese Second Office Action dated Sep. 19, 2017 issued in Application No. CN 201610622453.7. |
Chinese Third Office Action dated Jan. 24, 2018 issued in Application No. CN 201610622453.7. |
Chinese Fourth Office Action dated Jun. 1, 2018 issued in Application No. CN 201610622453.7. |
Chinese First Office Action dated Jul. 13, 2017 issued in Application No. CN 201610621114.7. |
Chinese Second Office Action dated Apr. 9, 2018 issued in Application No. CN 201610621114.7. |
Chinese Third Office Action dated Sep. 14, 2018 issued in Application No. CN 201610621114.7. |
Chinese First Office Action dated Jan. 14, 2019 issued in Application No. CN 201510117698.X. |
Chinese Second Office Action dated Jun. 21, 2019 issued in Application No. CN 201510117698.X. |
Chinese First Office Action dated May 13, 2020, issued in Application No. CN 201610377864.4. |
Chinese First Office Action dated Jan. 22, 2020, issued in Application No. CN 201780033558.1. |
European Extended Search Report dated Oct. 25, 2016 issued in Application No. EP 16 16 8661.3. |
European Office Action dated Mar. 19, 2019 issued in Application No. EP 16 16 8661.3. |
European Extended Search Report dated Sep. 9, 2019, issued in Application No. EP 17790575.9. |
International Search Report and Written Opinion—PCT/US2017/030190—ISA/US—dated Jul. 7, 2017 (Jul. 7, 2017). |
Litigation Document—“Complaint For Patent Infringement,” filed Sep. 3, 2015, in U.S. District Court of Delaware (Court Docket No. 1: 15-cv-00775-RGA). |
Litigation Document—“Report On the Filing or Determination of an Action Regarding a Patent or Trademark,” filed Sep. 3, 2015, in U.S. District Court of Delaware (Court Docket No. 1: 15-cv-00775-RGA). |
Litigation Document—“Complaint For Patent Infringement,” filed Oct. 29, 2015, in U.S. District Court of Delaware (Court Docket No. 1:15-cv-00990-RGA) [Re: U.S. Pat. Nos. 8,868,377, 8,920,332, and 9,089,760]. |
Litigation Document—“Report On the Filing or Determination of an Action Regarding a Patent or Trademark,” filed Oct. 29, 2015, in U.S. District Court of Delaware (Court Docket No. 1:15-cv-00990-RGA) [Re: U.S. Pat. Nos. 8,868,377, 8,920,332, and 9,089,760]. |
Litigation Document—“ORDER No. 24: Initial Determination Granting Respondents' Motion for Summary Determination of Invalidity under 35 U.S.C. § 101 with respect to all Three Asserted Patents and Terminating the Investigation in its Entirety,” filed Jul. 19, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof]. |
Litigation Document—“Respondents' Opposition to Complainant's Petition for Review of the Initial Determination Granting Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Aug. 8, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) (4446833v1/014972) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof]. |
Litigation Document—“Declaration of Majid Sarrafzadeh in Support of Complainant's Brief in Opposition to Respondents' Motion for Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Jun. 2, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 37-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof] [Exhibit 7]. |
Litigation Document—“Kiaei Declaration in Support of Complainant's Supplemental Brief Regarding Construction of “Operating the Heart Rate Monitor in a Worn Detection Mode” under 35 U.S.C. § 112(f),” filed Apr. 29, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 37-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof] [Exhibit 8]. |
Litigation Document—“Memorandum in Support of Respondents' Motion for Summary Deteninination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed May 23, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) (44325007v1/014972) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof]. |
Litigation Document—“Grimes Declaration in Support of Complainant's Brief in Opposition to Respondents' Motion for Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Jun. 2, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 37-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof] [Exhibit 28]. |
Litigation Document—“Complainant's Brief in Opposition to Respondents' Motion for Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Jun. 2, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof]. |
Litigation Document—“Complainant's Petition for Review of the Initial Determination Granting Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Aug. 1, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof]. |
Litigation Document—“Summary Pursuant to 19 C.F.R. § 210.43(b)(2) of Complainant's Petition for Review of the Initial Determination Granting Summary Determination that the Asserted Patents are Directed to Ineligible Subject Matter under 35 U.S.C. § 101,” filed Aug. 1, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof]. |
Litigation Document—“Notice of Commission Determination to Review an Initial Determination Granting Respondents' Motion for Summary Determination that Certain Asserted Claims are Directed to Ineligible Subject Matter under 35 U.S.C. § 101; and on Review to Remand the Investigation to the Presiding Administrative Law Judge,” issued Sep. 7, 2016, in United States International Trade Commission, Washington, D.C. (Investigation No. 337-TA-973) [In the Matter of Certain Wearable Activity Tracking Devices, Systems, and Components Thereof]. |
U.S. Appl. No. 61/736,310, filed Dec. 12, 2012, William Ahmed et al., entitled “Fitness Monitoring Systems and Methods Based on Continuous Collection of Physiological Data,” 61pp [Exhibit 4]. |
U.S. Appl. No. 61/696,525, filed Sep. 4, 2012, William Ahmed et al., entitled “Fitness Monitoring Systems and Methods Based on Continuous Collection of Physiological Data,” 47pp [Exhibit 5]. |
Gasparrini et al. (2013) “Evaluation and Possible Improvements of the ANT Protocol for Home Heart Monitoring Applications,” IEEE, 978-1-4673-2874-6/13, 7pp [Exhibit 6]. |
“UP3™, The world's most advanced tracker,” (Oct. 14, 2015) Jawbone, 10pp [Exhibit 12]. |
“UP4™, A fitness tracker so advanced it pays,” (Oct. 14, 2015) Jawbone, 12pp [Exhibit 13]. |
“User's Guide, MIO Drive+ Petite,” User's guide and how-to videos available at www.mioglobal.com, 3pp [Exhibit 16]. |
“SOLO 915, Heart Rate + Calorie Monitor,” (2009) SPORTLINE®, [retrieved on Oct. 15, 2010 at www.sportline.com] 25pp [Exhibit 17]. |
U.S. Notice of Allowance dated Oct. 14, 2014 issued in U.S. Appl. No. 14/295,144, 5pp [Exhibit 18]. |
“Health Touch™ Plus User Guide,” (2011) Timex Group USA, Inc., 12pp [Exhibit 18]. |
Czarnul, Pawel (Jun. 6-8, 2013) “Design of a Distributed System using Mobile Devices and Workflow Management for Measurement and Control of a Smart Home and Health,” Sopot, Poland, IEEE, pp. 184-192, 10pp [Exhibit 19]. |
Rabinovich, Roberto A., and Louvaris, Zafeiris et al. (Feb. 8, 2013) “Validity of Physical Activity Monitors During Daily Life in Patients With COPD,” ERJ Express, European Respiratory Society, 28pp [Exhibit 24]. |
Horvath et al. (2007) “The effect of pedometer position and normal gait asymmetry on step count accuracy,” Appl. Physiol. Nutr. Metab., 32:409-415, 8pp [Exhibit 32]. |
Graser et al. (2007) “Effects of Placement, Attachment, and Weight Classification on Pedometer Accuracy,” Journal of Physical Activity and Health, 4(4):359-369, 13pp [Exhibit 33]. |
Vyas et al. (2012) “Machine Learning and Sensor Fusion for Estimating Continuous Energy Expenditure,” AI Magazine, pp. 55-61, 13pp [Exhibit 42]. |
“New Lifestyles, NL-800 Activity Monitor, User's guide & record book,” (2005), New Lifestyles, Inc., 37pp. |
“StepWatch Step Activity Monitor, U.S. Pat. No. 5,485,402,” (2001) StepWatch™, Prosthetics Research Study, 7pp. |
Litigation Document—“Plaintiffs Original Complaint For Patent Infringement,” filed Jan. 4, 2016, in U.S. District Court for the Eastern District of North Carolina (Court Docket No. 5:16-cv-00002-FL) [Re: U.S Pat. Nos. 8,923,941, 8,886,269, 8,929,965 and 8,989,830], 11 pages. |
Cooper, Daniel (Aug. 16, 2013) Withings Pulse review, http://www.engadget.com/2013/08/16/withings-pulse-revew/, 8 pages. |
Dunn et al. (2007) “A Novel Ring Shaped Photodiode for Reflectance Pulse Oximetry in Wireless Applications,” IEEE Sensors Conference, pp. 596-599. |
Kim, D. et al. A Linear Transformation Approach for Estimating Pulse Arrival Time. Journal of Applied Mathematics. vol. 2012. Jan. 20, 2012. [Retrieve Jun. 19, 2017]. Retrieved from internet: <https://www.emis.de/journals/HOA/JAM/Volume2012/643653.pdf> pp. 1-12. |
LIFETRNR, User Manual (2003, specific date unknown), NB new balance®, Implus Footcare, LLC, 3 pages. |
Rainmaker, (Jun. 25, 2012, updated Feb. 16, 2013) “Garmin Swim watch In-Depth Review,” [retrieved on Sep. 9, 2013 at http://www.dcrainmaker.com/2012/06/garmin-swim-in-depth-review.html, 38 pp. |
“Withings pulse, Quick Installation Guide” (Jul. 24, 2013) Withings Pulse QIG, v 1.3, withings.com/pulse, 16 pages. |
Zijlstra, Wiebren, (2004) “Assessment of spatio-temporal parameters during unconstrained walking,” Eur J Appl Physiol, 92:39-44. |
U.S. Appl. No. 14/214,655, filed Mar. 14, 2014, Hong et al. |
U.S. Appl. No. 15/494,257, filed Apr. 21, 2017, Richards et al. |
U.S. Appl. No. 15/948,970, filed Apr. 9, 2018, Nadeau et al. |
U.S. Appl. No. 16/592,599, filed Oct. 3, 2019, Weekly et al. |
U.S. Appl. No. 16/798,257, filed Feb. 21, 2020, Wu et al. |
Number | Date | Country | |
---|---|---|---|
20180092551 A1 | Apr 2018 | US |
Number | Date | Country | |
---|---|---|---|
61924547 | Jan 2014 | US | |
61752826 | Jan 2013 | US | |
61830600 | Jun 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14639409 | Mar 2015 | US |
Child | 15827970 | US | |
Parent | 14302360 | Jun 2014 | US |
Child | 14639409 | US | |
Parent | 14156381 | Jan 2014 | US |
Child | 14302360 | US |