The present invention relates generally to health and environmental monitors and, more particularly, to wireless health and environment monitors.
There is growing market demand for personal health and environmental monitors, for example, for gauging overall health and metabolism during exercise, athletic training, dieting, and physical therapy. However, traditional health monitors and environmental monitors may be bulky, rigid, and uncomfortable—generally not suitable for use during daily physical activity. There is also growing interest in generating and comparing health and environmental exposure statistics of the general public and particular demographic groups. For example, collective statistics enable the healthcare industry and medical community to direct healthcare resources to where they are most highly valued. However, methods of collecting these statistics may be expensive and laborious, often utilizing human-based recording/analysis steps at multiple sites.
As such, improved ways of collecting, storing and analyzing personal health and environmental information are needed. In addition, improved ways of distributing raw and analyzed personal health and environmental information are desirable to support efforts to enhance healthcare quality and reduce costs.
In view of the above discussion, apparatus for monitoring various physiological and environmental factors are provided. According to some embodiments of the present invention, real-time, noninvasive health and environmental monitors include a plurality of compact sensors integrated within small, low-profile devices. Physiological and environmental data is collected and wirelessly transmitted into a wireless network, where the data is stored and/or processed.
In some embodiments of the invention, an earpiece functions as a physiological monitor, an environmental monitor, and a wireless personal communicator. The earpiece can take advantage of commercially available open-architecture wireless paradigms, such as Bluetooth®, Wi-Fi, or ZigBee. In some embodiments, a small, compact earpiece contains at least one microphone and one speaker, and is configured to transmit information wirelessly to a recording device such as, for example, a cell phone, a personal digital assistant (PDA), and/or a computer. The earpiece contains a plurality of sensors for monitoring personal health and environmental exposure. Health and environmental information, sensed by the sensors is transmitted wirelessly, in real-time, to a recording device, capable of processing and organizing the data into meaningful displays, such as charts. In some embodiments, an earpiece user can monitor health and environmental exposure data in real-time, and may also access records of collected data throughout the day, week, month, etc., by observing charts and data through an audio-visual display.
In some embodiments, an earpiece can integrate personal physiological and environmental exposure information with biofeedback and personal entertainment. In other embodiments of the present invention, earpiece monitor devices enable a variety of networks, applications, games, and business methods.
In some embodiments of the present invention, a monitoring apparatus includes a housing configured to be attached to the body of a person, one or more physiological sensors and one or more environmental sensors supported by (within and/or on) the housing. Each physiological sensor is configured to detect and/or measure physiological information from the person, and each environmental sensor is configured to detect and/or measure environmental conditions in a vicinity of the person wearing the apparatus. The apparatus also includes a signal processor that is configured to receive and process signals produced by the physiological and environmental sensors. A wireless transmitter is responsive to the signal processor and is configured to wirelessly transmit physiological and environmental sensor signals as processed by the signal processor from the signal processor to a remote terminal in real-time.
Each physiological sensor is configured to detect and/or measure one or more of the following types of physiological information: heart rate, pulse rate, breathing rate, blood flow, heartbeat signatures, cardio-pulmonary health, organ health, metabolism, electrolyte type and/or concentration, physical activity, caloric intake, caloric metabolism, blood metabolite levels or ratios, blood pH level, physical and/or psychological stress levels and/or stress level indicators, drug dosage and/or dosimetry, physiological drug reactions, drug chemistry, biochemistry, position and/or balance, body strain, neurological functioning, brain activity, brain waves, blood pressure, cranial pressure, hydration level, auscultatory information, auscultatory signals associated with pregnancy, physiological response to infection, skin and/or core body temperature, eye muscle movement, blood volume, inhaled and/or exhaled breath volume, physical exertion, exhaled breath physical and/or chemical composition, the presence and/or identity and/or concentration of viruses and/or bacteria, foreign matter in the body, internal toxins, heavy metals in the body, anxiety, fertility, ovulation, sex hormones, psychological mood, sleep patterns, hunger and/or thirst, hormone type and/or concentration, cholesterol, lipids, blood panel, bone density, organ and/or body weight, reflex response, sexual arousal, mental and/or physical alertness, sleepiness, auscultatory information, response to external stimuli, swallowing volume, swallowing rate, sickness, voice characteristics, voice tone, voice pitch, voice volume, vital signs, head tilt, allergic reactions, inflammation response, auto-immune response, mutagenic response, DNA, proteins, protein levels in the blood, water content of the blood, pheromones, internal body sounds, digestive system functioning, cellular regeneration response, healing response, stem cell regeneration response
Each environmental sensor is configured to detect and/or measure one or more of the following types of environmental information: climate, humidity, temperature, pressure, barometric pressure, soot density, airborne particle density, airborne particle size, airborne particle shape, airborne particle identity, volatile organic chemicals (VOCs), hydrocarbons, polycyclic aromatic hydrocarbons (PAHs), carcinogens, toxins, electromagnetic energy, optical radiation, X-rays, gamma rays, microwave radiation, terahertz radiation, ultraviolet radiation, infrared radiation, radio waves, atomic energy alpha particles, atomic energy beta-particles, gravity, light intensity, light frequency, light flicker, light phase, ozone, carbon monoxide, carbon dioxide, nitrous oxide, sulfides, airborne pollution, foreign material in the air, viruses, bacteria, signatures from chemical weapons, wind, air turbulence, sound and/or acoustical energy, ultrasonic energy, noise pollution, human voices, animal sounds, diseases expelled from others, exhaled breath and/or breath constituents of others, toxins from others, pheromones from others, industrial and/or transportation sounds, allergens, animal hair, pollen, exhaust from engines, vapors and/or fumes, fuel, signatures for mineral deposits and/or oil deposits, snow, rain, thermal energy, hot surfaces, hot gases, solar energy, hail, ice, vibrations, traffic, the number of people in a vicinity of the person, coughing and/or sneezing sounds from people in the vicinity of the person, loudness and/or pitch from those speaking in the vicinity of the person.
In some embodiments, the signal processor is configured to process signals produced by the physiological and environmental sensors into signals that can be heard and/or viewed by the person wearing the apparatus. In some embodiments, the signal processor is configured to selectively extract environmental effects from signals produced by a physiological sensor and/or selectively extract physiological effects from signals produced by an environmental sensor.
In some embodiments of the present invention, a monitoring apparatus configured to be worn by a person includes a physiological sensor that is oriented in a direction towards the person and an environmental sensor that is oriented in a direction away from the person. A buffer material is positioned between the physiological sensor and environmental sensors and is configured to selectively reflect and/or absorb energy emanating from the environment and/or the person.
In some embodiments of the present invention, a monitoring apparatus may include a receiver that is configured to receive audio and/or video information from a remote terminal, and a communication module that is configured to store and/or process and/or play audio and/or video information received from the remote terminal. In some embodiments, the communication module may be configured to alert (e.g., via audible and/or visible and/or physical alerts) a person wearing the apparatus when a physiological sensor detects certain physiological information from the person and/or when an environmental sensor detects certain environmental information from the vicinity of the person. In some embodiments, the communication module is configured to audibly present vital sign information to the person wearing the apparatus. In some embodiments, the communication module may be configured to store content generated by the person.
In some embodiments of the present invention, a monitoring apparatus may include a transmitter that is configured to transmit signals produced by physiological and environmental sensors associated therewith to a gaming device. The monitoring apparatus may also be configured to receive feedback regarding monitored health and environmental parameters. As such, personal health and environmental feedback can be an active component of a game.
In some embodiments, the apparatus is an earpiece module that is configured to be attached to the ear of a person, and includes a speaker, microphone, and transceiver that is electronically connected to the speaker and microphone and that permits bidirectional wireless communications between the earpiece module and a remote terminal, such as a cell phone. The transceiver (e.g., a Bluetooth®, Wi-Fi, or ZigBee transceiver) is electronically connected to the signal processor and is configured to transmit physiological and environmental sensor signals from the signal processor to the remote terminal. In some embodiments, the earpiece module may include an arm that is attached to the housing and that supports the microphone. The arm may be movable between a stored position and an extended, operative position. The arm may also include one or more physiological sensor and/or environmental sensors.
In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes a first acoustical sensor oriented in a direction towards a tympanic membrane of the ear and is configured to detect acoustical energy emanating from the tympanic membrane. A second acoustical sensor is oriented in a direction away from the person. The signal processor is configured to utilize signals produced by the second acoustical signal to extract environmental acoustical energy not emanating from the tympanic membrane from signals produced by the first acoustical sensor. In some embodiments, the earpiece module may include an optical emitter that directs optical energy towards the tympanic membrane, and an optical detector that is configured to detect secondary optical energy emanating from the tympanic membrane. The signal processor is configured to extract selected optical energy from the secondary optical energy emanating from the tympanic membrane. The signal processor may also be configured to extract optical noise from the secondary optical energy emanating from the tympanic membrane. In some embodiments, the optical detector may include a filter configured to pass secondary optical energy at selective wavelengths.
In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes an optical detector that is configured to detect acoustically modulated blackbody IR radiation emanating from the tympanic membrane.
In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes an optical emitter that directs optical energy towards the tympanic membrane, and an optical detector configured to detect secondary optical energy emanating from the tympanic membrane. In some embodiments, the signal processor may be configured to extract selected optical energy and/or optical noise from the secondary optical energy emanating from the tympanic membrane. In some embodiments, the optical detector may include a filter configured to pass secondary optical energy at selective wavelengths.
In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes an ear hook that is configured to attach to an ear of a person. One or more physiological sensors and/or one or more environmental sensors may be supported by the ear hook. In some embodiments, the hook may include a pinna cover that is configured to contact a portion of the pinna of an ear. One or more physiological and/or environmental sensors may be supported by the pinna cover.
In some embodiments of the present invention, an earpiece module may include an arm that extends outwardly therefrom and that supports one or more physiological sensors and/or environmental sensors. For example, the arm may be configured to support physiological sensors configured to detect and/or measure jaw motion and/or arterial blood flow near the neck of a person wearing the earpiece module.
In some embodiments of the present invention, an earpiece module may include an earpiece fitting configured to be inserted near or within the ear canal of a person wearing the earpiece. The earpiece fitting may include one or more physiological sensors configured to detect information from within the ear canal.
In some embodiments of the present invention, an earpiece module may include a transmittance pulse oximeter and/or reflectance pulse oximeter. For example, the earpiece module may include an earlobe clip having a transmittance pulse oximeter and/or reflectance pulse oximeter supported thereby. As another example, the earpiece module may include a transmitter pulse oximeter and/or reflectance pulse oximeter supported at the front or back of the ear.
In some embodiments of the present invention, a monitoring apparatus is an earring. The earring may be configured to operate independently of other monitoring apparatus, such as an earpiece module, or may operate in conjunction with another monitoring apparatus. For example, an earring may include one or more physiological sensors configured to detect and/or measure physiological information from the person, and one or more environmental sensors configured to detect and/or measure environmental conditions in a vicinity of the person wearing the earring. The earring may also include a signal processor that receives and processes signals produced by the physiological and environmental sensors, and a transmitter that transmits physiological and environmental sensor signals from the signal processor to a remote terminal in real-time.
In some embodiments of the present invention, a monitoring apparatus configured to be attached to the ear of a person may include a housing containing one or more physiological and environmental sensors wherein the housing is configured to be positioned in adjacent contacting relationship with the temple of the person.
Monitoring apparatus, according to some embodiments of the present invention, may include various additional devices/features. For example, a monitoring apparatus may include an air sampling system that samples air in a vicinity of the person wearing the apparatus. In some embodiments, one or more physiological sensors in a monitoring apparatus may be configured to detect drowsiness of the person wearing the apparatus. An alarm may be provided that is configured to alert the person in response to one or more physiological sensors detecting drowsiness. In some embodiments, a monitoring apparatus may include a user interface that provides user control over one or more of the physiological and/or environmental sensors. A user interface may be provided on the monitoring apparatus or may be included on a remote device in wireless communication with the monitoring apparatus. In some embodiments, a monitoring apparatus may include a user interface that is configured to allow the person to store a time mark indicating a particular point in time.
Monitoring apparatus, according to some embodiments of the present invention, may be configured to send a signal to a remote terminal when one or more of the physiological and/or environmental sensors are turned off by a user and/or when one or more of the physiological and/or environmental sensors malfunction or fail. In some embodiments, a signal may be sent to a remote terminal when potentially erroneous data has been collected by one or more of the physiological and/or environmental sensors, such as when a person wearing a monitoring apparatus is surrounded by loud noises.
Monitoring apparatus, according to some embodiments of the present invention, may be configured to detect damage to a portion of the body of the person wearing the apparatus, and may be configured to alert the person when such damage is detected. For example, when a person is exposed to sound above a certain level that may be potentially damaging, the person is notified by the apparatus to move away from the noise source. As another example, the person may be alerted upon damage to the tympanic membrane due to loud external noises.
The present invention now is described more fully hereinafter with reference to the accompanying drawings, in which preferred embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art.
Like numbers refer to like elements throughout. In the figures, the thickness of certain lines, layers, components, elements or features may be exaggerated for clarity.
The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.
Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the specification and relevant art and should not be interpreted in an idealized or overly formal sense unless expressly so defined herein. Well-known functions or constructions may not be described in detail for brevity and/or clarity.
It will be understood that when an element is referred to as being “on”, “attached” to, “connected” to, “coupled” with, “contacting”, etc., another element, it can be directly on, attached to, connected to, coupled with or contacting the other element or intervening elements may also be present. In contrast, when an element is referred to as being, for example, “directly on”, “directly attached” to, “directly connected” to, “directly coupled” with or “directly contacting” another element, there are no intervening elements present. It will also be appreciated by those of skill in the art that references to a structure or feature that is disposed “adjacent” another feature may have portions that overlap or underlie the adjacent feature.
Spatially relative terms, such as “under”, “below”, “lower”, “over”, “upper” and the like, may be used herein for ease of description to describe one element or feature's relationship to another element(s) or feature(s) as illustrated in the figures. It will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures. For example, if the device in the figures is inverted, elements described as “under” or “beneath” other elements or features would then be oriented “over” the other elements or features. Thus, the exemplary term “under” can encompass both an orientation of “over” and “under”. The device may be otherwise oriented (rotated 90 degrees or at other orientations) and the spatially relative descriptors used herein interpreted accordingly. Similarly, the terms “upwardly”, “downwardly”, “vertical”, “horizontal” and the like are used herein for the purpose of explanation only unless specifically indicated otherwise.
The term “earpiece module” includes any type of device that may be attached to or near the ear of a user and may have various configurations, without limitation.
The term “real-time” is used to describe a process of sensing, processing, or transmitting information in a time frame which is equal to or shorter than the minimum timescale at which the information is needed. For example, the real-time monitoring of pulse rate may result in a single average pulse-rate measurement every minute, averaged over 30 seconds, because an instantaneous pulse rate is often useless to the end user. Typically, averaged physiological and environmental information is more relevant than instantaneous changes. Thus, in the context of the present invention, signals may sometimes be processed over several seconds, or even minutes, in order to generate a “real-time” response.
The term “monitoring” refers to the act of measuring, quantifying, qualifying, estimating, sensing, calculating, interpolating, extrapolating, inferring, deducing, or any combination of these actions. More generally, “monitoring” refers to a way of getting information via one or more sensing elements. For example, “blood health monitoring” includes monitoring blood gas levels, blood hydration, and metabolite/electrolyte levels.
The term “physiological” refers to matter or energy of or from the body of a creature (e.g., humans, animals, etc.). In embodiments of the present invention, the term “physiological” is intended to be used broadly, covering both physical and psychological matter and energy of or from the body of an organism. However, in some cases, the term “psychological” is called-out separately to emphasize aspects of physiology that are more closely tied to conscious or subconscious brain activity rather than the activity of other organs, tissues, or cells.
The term “psychosocial stress” refers to events of psychological or social origin which challenge the homeostatic state of biological systems.
The term “body” refers to the body of a person (or animal) that may utilize an earpiece module according to embodiments of the present invention. Monitoring apparatus, according to embodiments of the present invention may be worn by humans and animals.
In the following figures, earpiece modules will be illustrated and described for attachment to the ear of the human body. However, it is to be understood that embodiments of the present invention are not limited to those worn by humans. Moreover, monitoring apparatus according to embodiments of the present invention are not limited to earpiece modules and/or devices configured to be attached to or near the ear. Monitoring apparatus according to embodiments of the present invention may be worn on various parts of the body.
Some embodiments of the present invention may arise from a discovery that the ear is an ideal location on the human body for a wearable health and environmental monitor. The ear is a relatively immobile platform that does not obstruct a person's movement or vision. Devices located along the ear have access to the inner-ear canal and tympanic membrane (for measuring core body temperature), muscle tissue (for monitoring muscle tension), the pinna and earlobe (for monitoring blood gas levels), the region behind the ear (for measuring skin temperature and galvanic skin response), and the internal carotid artery (for measuring cardiopulmonary functioning). The ear is also at or near the point of exposure to: environmental breathable toxicants of interest (volatile organic compounds, pollution, etc.; noise pollution experienced by the ear; and lighting conditions for the eye. Furthermore, as the ear canal is naturally designed for transmitting acoustical energy, the ear provides an optimal location for monitoring internal sounds, such as heartbeat, breathing rate, and mouth motion.
Bluetooth-enabled and/or other personal communication earpiece modules may be configured to incorporate physiological and/or environmental sensors, according to some embodiments of the present invention. Bluetooth earpiece modules are typically lightweight, unobtrusive devices that have become widely accepted socially. Moreover, Bluetooth earpiece modules are cost effective, easy to use, and are often worn by users for most of their waking hours while attending or waiting for cell phone calls. Bluetooth earpiece modules configured according to embodiments of the present invention are advantageous because they provide a function for the user beyond health monitoring, such as personal communication and multimedia applications, thereby encouraging user compliance. Exemplary physiological and environmental sensors that may be incorporated into a Bluetooth or other type of earpiece module include, but are not limited to accelerometers, auscultatory sensors, pressure sensors, humidity sensors, color sensors, light intensity sensors, pressure sensors, etc.
Wireless earpiece devices incorporating low-profile sensors and other electronics, according to embodiments of the present invention, offer a platform for performing near-real-time personal health and environmental monitoring in wearable, socially acceptable devices. The capability to unobtrusively monitor an individual's physiology and/or environment, combined with improved user compliance, is expected to have significant impact on future planned health and environmental exposure studies. This is especially true for those that seek to link environmental stressors with personal stress level indicators. The large scale commercial availability of this low-cost device can enable cost-effective large scale studies. The combination of monitored data with user location via GPS data can make on-going geographic studies possible, including the tracking of infection over large geographic areas. The commercial application of the proposed platform encourages individual-driven health maintenance and promotes a healthier lifestyle through proper caloric intake and exercise.
Accordingly, some embodiments of the present invention combine a personal communications earpiece device with one or more physiological and/or environmental sensor. Other embodiments may combine physiological and/or environmental sensors into an earpiece device.
Embodiments of the present invention are not limited to devices that communicate wirelessly. In some embodiments of the present invention, devices configured to monitor an individual's physiology and/or environment may be wired to a device that stores and/or processes data. In some embodiments, this information may be stored on the earpiece module itself.
A physiological sensor 101 can be any compact sensor for monitoring the physiological functioning of the body, such as, but not limited to, sensors for monitoring: heart rate, pulse rate, breathing rate, blood flow, heartbeat signatures, cardio-pulmonary health, organ health, metabolism, electrolyte type and concentration, physical activity, caloric intake, caloric metabolism, metabolomics, physical and psychological stress levels and stress level indicators, physiological and psychological response to therapy, drug dosage and activity (drug dosimetry), physiological drug reactions, drug chemistry in the body, biochemistry, position & balance, body strain, neurological functioning, brain activity, brain waves, blood pressure, cranial pressure, hydration level, auscultatory information, auscultatory signals associated with pregnancy, physiological response to infection, skin and core body temperature, eye muscle movement, blood volume, inhaled and exhaled breath volume, physical exertion, exhaled breath physical and chemical composition, the presence, identity, and concentration of viruses & bacteria, foreign matter in the body, internal toxins, heavy metals in the body, anxiety, fertility, ovulation, sex hormones, psychological mood, sleep patterns, hunger & thirst, hormone type and concentration, cholesterol, lipids, blood panel, bone density, body fat density, muscle density, organ and body weight, reflex response, sexual arousal, mental and physical alertness, sleepiness, auscultatory information, response to external stimuli, swallowing volume, swallowing rate, sickness, voice characteristics, tone, pitch, and volume of the voice, vital signs, head tilt, allergic reactions, inflammation response, auto-immune response, mutagenic response, DNA, proteins, protein levels in the blood, body hydration, water content of the blood, pheromones, internal body sounds, digestive system functioning, cellular regeneration response, healing response, stem cell regeneration response, and the like. Vital signs can include pulse rate, breathing rate, blood pressure, pulse signature, body temperature, hydration level, skin temperature, and the like. A physiological sensor may include an impedance plethysmograph for measuring changes in volume within an organ or body (usually resulting from fluctuations in the amount of blood or air it contains). For example, the earpiece module 100 may include an impedance plethysmograph to monitor blood pressure in real-time.
An external energy sensor 102, serving primarily as an environmental sensor, can be any compact sensor for monitoring the external environment in the vicinity of the body, such as, but not limited to, sensors for monitoring: climate, humidity, temperature, pressure, barometric pressure, pollution, automobile exhaust, soot density, airborne particle density, airborne particle size, airborne particle shape, airborne particle identity, volatile organic chemicals (VOCs), hydrocarbons, polycyclic aromatic hydrocarbons (PAHs), carcinogens, toxins, electromagnetic energy (optical radiation, X-rays, gamma rays, microwave radiation, terahertz radiation, ultraviolet radiation, infrared radiation, radio waves, and the like), EMF energy, atomic energy (alpha particles, beta-particles, gamma rays, and the like), gravity, light properties (such as intensity, frequency, flicker, and phase), ozone, carbon monoxide, greenhouse gases, CO2, nitrous oxide, sulfides, airborne pollution, foreign material in the air, biological particles (viruses, bacteria, and toxins), signatures from chemical weapons, wind, air turbulence, sound and acoustical energy (both human audible and inaudible), ultrasonic energy, noise pollution, human voices, animal sounds, diseases expelled from others, the exhaled breath and breath constituents of others, toxins from others, bacteria & viruses from others, pheromones from others, industrial and transportation sounds, allergens, animal hair, pollen, exhaust from engines, vapors & fumes, fuel, signatures for mineral deposits or oil deposits, snow, rain, thermal energy, hot surfaces, hot gases, solar energy, hail, ice, vibrations, traffic, the number of people in a vicinity of the user, the number of people encountered throughout the day, other earpiece module users in the vicinity of the earpiece module user, coughing and sneezing sounds from people in the vicinity of the user, loudness and pitch from those speaking in the vicinity of the user, and the like.
In some embodiments, a physiological sensor 101 and/or an environmental sensor 102 may be configured to identify a person to whom the earpiece module 100 is attached.
In some embodiments, a physiological sensor 101 and/or an environmental sensor 102 may be configured to monitor physical aging rate of a person or subject. The signal processor 103 may be configured to processes information from a physiological sensor and/or an environmental sensor to assess aging rate. Physiological sensors configured to assess aging rate may include pulse rate sensors, blood pressure sensors, activity sensors, and psychosocial stress sensors. Environmental sensors configured to assess aging rate may include UV sensors and pollution sensors.
In some embodiments, a physiological sensor 101 and/or an environmental sensor 102 may be configured to be regenerated through a physical and/or chemical change. For example, it is anticipated that an earpiece module 100, or other device incorporating physiological and/or environmental sensors according to embodiments of the present invention may be coupled to an apparatus that is configured to “recharge” or regenerate one or more environmental and/or physiological sensors via a physical process or a chemical process, etc.
Because the earpiece module is capable of measuring and transmitting sensor information in real-time over a duration of time, the physiological and environmental sensors 101, 102 can be used to sense the aforementioned parameters over time, enabling a time-dependent analysis of the user's health and environment as well as enabling a comparison between the user's health and environment. Combined with proximity or location detection, this allows an analysis for pinpointing the location where environmental stress and physical strain took place. The signal processor 103 provides a means of converting the digital or analog signals from the sensors 101, 102 into data that can be transmitted wirelessly by the transmitter 104. The signal processor 103 may be composed of, for example, signal conditioners, amplifiers, filters, digital-to-analog and analog-to-digital converters, digital encoders, modulators, mixers, multiplexers, transistors, various switches, microprocessors, or the like. For personal communication, the signal processor 103 processes signals received by the receiver 104 into signals that can be heard or viewed by the user. The received signals may also contain protocol information for linking various telemetric modules together, and this protocol information can also be processed by the signal processor 103. The signal processor 103 may utilize one or more “compression/decompression algorithms used in digital media” (CODECs) for processing data. The transmitter 104 can be comprised of a variety of compact electromagnetic transmitters. A standard compact antenna is used in the standard Bluetooth headset protocol, but any kind of electromagnetic antenna suitable for transmitting at human-safe electromagnetic frequencies may be utilized. The receiver 104 can also be an antenna. In some embodiments, the receiving antenna and the transmitting antenna are physically the same. The receiver/transmitter 104 can be, for example, a non-line-of-sight (N LOS) optical scatter transmission system. These systems typically use short-wave (blue or UV) optical radiation or “solar blind” (deep-UV) radiation in order to promote optical scatter, but IR wavelengths can also suffice. Additionally, a sonic or ultrasonic transmitter can be used as the receiver/transmitter 104 of the earpiece module 100, but preferably using sounds that are higher or lower than the human hearing range. A variety of sonic and ultrasonic receivers and transmitters are available in the marketplace and may be utilized in accordance with embodiments of the present invention. If a telecommunication device 210 (
In some embodiments, the transmitter/receiver 104 is configured to transmit signals from the signal processor to the remote terminal following a predetermined time interval. For example, the transmitter may delay transmission until a certain amount of detection time has elapsed, until a certain amount of processing time has elapsed, etc.
The power source can be any portable power source 106 capable of fitting inside the earpiece module housing. According to some embodiments, the power source 106 is a portable rechargeable lithium-polymer or zinc-air battery. Additionally, portable energy-harvesting power sources can be integrated into the earpiece module 100 and can serve as a primary or secondary power source. For example, a solar cell module can be integrated into the earpiece module 100 for collecting and storing solar energy. Additionally, piezoelectric devices or microelectromechanical systems (MEMS) can be used to collect and store energy from body movements, electromagnetic energy, and other forms of energy in the environment or from the user himself. A thermoelectric or thermovoltaic device can be used to supply some degree of power from thermal energy or temperature gradients. In some embodiments, a cranking or winding mechanism can be used to store mechanical energy for electrical conversion or to convert mechanical energy into electrical energy that can be used immediately or stored for later.
The various components describe above are configured to fit within the earpiece housing 108 and/or be attached thereto. The earpiece module housing 108 may be formed from any safe and comfortable solid material, such as metal, rubber, wood, polymers, ceramic, organic materials, or various forms of plastic. The earpiece attachment component 105 is attached to the earpiece module housing 108 and is designed to fit around or near the ear. For example, the standard Bluetooth headset includes an earpiece attachment that is connected to the headset housing via a double-jointed socket, to provide comfort and positioning flexibility for the user. In some embodiments, the earpiece attachment component 105 can be part of the housing 108, such that the entire earpiece module is one largely inflexible, rigid unit. In such case, a counterweight may be incorporated into the earpiece module 100 to balance the weight of the earpiece electronics and power source. In some embodiments, the earpiece attachment component 105 can contain physiological and environmental sensors, and the earpiece attachment component 105 may be detachable. In some embodiments, more than one earpiece attachment 105 can be attached to the earpiece module housing 108.
The communication and entertainment module 107 (also interchangeably referred to as a “communication module”) is used for, but not limited to: processing or generating an audible sound from information received via the receiver 104 (from a cell phone, computer, network, database, or the like) and/or processing or generating an electrical signal from an audible sound from the user such that the electrical signal can be transmitted telemetrically via the transmitter 104. For example, in standard Bluetooth protocol, communication electronics are used to convert an audible conversation into an electrical signal for telemetric conversation; communication electronics are also used to convert a digitized telemetric conversation into an audible conversation for the earpiece user. Additionally, the communication and entertainment module 107 can be used to store, process, or play analog or digital information from music, radio shows, videos, or other audible entertainment and to communicate this information to an earpiece user. In many cases, this information includes information received by the receiver 104. In many cases, the analog or digital information is not stored in the communication and entertainment module 107 but, rather, is stored in a portable telecommunication device 210 (
In some embodiments, an audible communicator is provided that is configured to communicate therapeutic sounds (e.g., music therapy, etc.) to the person in response to physiological or psychosocial stress. The audible communicator may be embodied in the communication and entertainment module 107 or may be a separate speaker. In some embodiments, light therapy may be provided to the person in response to physiological or psychosocial stress. In some embodiments, the communication and entertainment module 107 may be configured to communicate a treatment, therapy, and/or plan of action to the person upon detection of physiological and/or environmental concerns. For example, if it is detected that the person is being exposed to unhealthy doses of UV radiation, the communication and entertainment module 107 may audibly instruct the person to move away from the person's current location (e.g., move indoors, etc.).
Like the other components of the earpiece module 100 shown in
Bluetooth devices conventionally contain a communication module, such as communication module 107, for converting digital or analog information into audible sounds for the user. However, when combined with the health and environmental monitoring properties of an earpiece module 100 according to embodiments of the present invention, the communication and entertainment module 107 can provide functionality. For example, the earpiece module can serve as a biofeedback device. As a non-limiting example, if a user is in a polluted environment, such as air filled with VOCs, the communication module 107 may notify the user to move to a new environment. As another example, if one or more of the physiological and environmental sensors 101, 102 of the earpiece module 100 pick up a high particulate density in the environment, with an elevation in core body temperature, and a change in voice pitch occurring simultaneously (or near-simultaneously) within a common timeframe, the communication module 107 may alert the user that he/she may be having an allergic response. As a further example, the user can use the communication and entertainment module 107 to execute biofeedback for willfully controlling blood pressure, breathing rate, body temperature, pulse rate, and the like. The communication module 107 may utilize audible or visible alerts if the user is meeting their physiological targets or exceeding safe physiological limits. Alerting a user by physical or electrical force, such as the sense of touch or tingling from an electric pulse or vibration, can also be utilized. Thus, although communication by audible means is often utilized, the communication module 107 can alert, signify, or communicate with the user through sound, light, electrical actuation, and physical actuation.
As a second example of this biofeedback method, basic vital signs collected by the physiological sensors 101 and processed by the signal processor 103 can be presented to the earpiece user audibly, through the communication and entertainment module 107. For example, the user may be able to listen to his/her breathing rate, pulse rate, and the like. Additionally, an entertaining or aggravating sound or song can be used to alert the user to favorable or unfavorable personal health and environmental factors occurring in real-time. This technique may be applied towards education, such as positive or negative feedback for educational games, learning games, or games of deception (e.g., poker, etc.).
In some embodiments, the earpiece module 100 may be configured to deliver and/or monitor drugs. For example, a transdermal drug delivery system may be provided that is controlled by earpiece electronics. Earpiece sensors can monitor the drug dosage and the physiological effects of the drug in real-time.
A health and environmental network 200 according to embodiments of the present invention that may incorporate the earpiece module 100 of
The sensor module 201 can be composed of a primary module alone or a primary module and at least one secondary module. The secondary modules can be located at any location of the body, but are preferably located in a region near the ear, and preferably the earpiece module 100 serves as the primary module. In most cases, the secondary modules are not necessary. But in some cases, secondary modules may be located, for example, behind the ear (near the lymph nodes), at or near the earlobes (such as one or more earrings or ear clips), at the front of the ear (near the carotid artery), at the temples, along the neck, or other locations near the ear. These wearable secondary modules can be connected with either a “hard” connection to the primary module (such as an electric cable) or a “soft” connection to the primary module (such as a wireless connection). In Bluetooth protocol, each secondary module can be simultaneously in direct wireless communication with the primary module. Primary modules and secondary modules in the same location can promote quick-donning, ease-of-use, and comfortability for the end user. Users are not prone to accept multiple modules at multiple locations of the body.
The earpiece sensor module 201 communicates wirelessly with the portable telecommunication device 210, preferably in an open architecture configuration, such as Bluetooth or ZigBee. The telecommunication device 210 can be any portable device, such as a cell phone, PDA, laptop computer, Blackberry, another earpiece, or other portable, telemetric device. The portable telecommunication device 210 and the earpiece module 201 can telemetrically communicate both to and from each other. Though the main purpose of the portable telecommunication device is to transmit the local wireless signal from the sensor module 101 over longer distances unattainable by the transmitter 104 of the sensor module 201, the telecommunication 210 can also serve as a method of personal communication and entertainment for the earpiece user.
In some embodiments, the telecommunication device 210 transmits data in only one direction or particular directions. For example, in one embodiment, the portable telecommunication device 210 can receive telemetric information from the sensor module 201 but cannot send out signals to a transmission system 211. The portable telecommunication device 210 may also contain an end-user graphical interface, such as a user interface 214 in the network 200, such that data from the earpiece module 201 can be stored, analyzed, summarized, and displayed on the portable telecommunication device 210. For example, charts relating health and environment, as well as real-time biofeedback and the like, can be displayed on a cell phone, media player, PDA, laptop, or other device. The telecommunication device 210 may also contain physiological and environmental sensors itself, such as blood pressure, pulse rate, and pulse-oximetry, and the like. Additionally, the telecommunication device 210 can communicate with the earpiece module 201 to transfer commands, activate or deactivate sensors, communicate with the user, and the like.
The portable telecommunication device 210 sends/receives wireless information directly to/from a transmission system 211 for transmission to a database (such as personal database 312 and/or anonymous database 313) for storage, analysis, and retrieval of data. The style of transmission system depends largely on the location of the database. For example, if the database is located in a local computer, the wireless information from the telecommunication device 210 can be sent directly to the local computer. This computer may be connected with the Internet, allowing access to the database from the web. However, the database is more typically located far away from the user and telecommunication module. In this case, the wireless signal from the telecommunication device 210 can be sent to a reception tower and routed through a base station. This information can then be sent to a database through the Internet. A variety of other transmission protocols can be applied for connection between the telecommunication device 210 and the databases 212, 213.
The personal and anonymous databases 212, 213 represent databases that may or may not be located on the same computer. A key difference between these two databases is not the physical location of the database but rather the type of information available on each database. For example, the anonymous database 213, containing health and environmental data from multiple indistinct earpiece users, can be public and accessible through the Internet by various users. In contrast, the personal database 212 contains health and environmental data that is personalized for each user, including personalized information such as name, birth date, address, and the like. Users can log-in to their personalized information in the personal database 212 through an interactive user interface 214 and compare this information with information from multiple users in the anonymous database 213 via a graphical user interface.
The user interface 214 can be a computer monitor, a cell phone monitor, a PDA monitor, a television, a projection monitor, a visual monitor on the earpiece module 201, or any method of visual display. (Audible methods and audio-visual methods can also be used for the user interface 214.) For example, the user may log-in to their personal database 212 through a computer user interface 214 and compare real-time personal health and environmental exposure data with that of other users on the network 200. In some cases, the data from other users may be anonymous statistics. In some cases, one or more users may have agreements to view the data of one or more other users, and in other cases, users may agree to share mutual personalized data through the Internet.
A specific embodiment of a graphical user interface 300 is presented in
The network 200 can be used in medicine for a variety of important functions. As one example, a doctor can monitor the health of patients through each patient's personalized database 212. If the earpiece module 201 contains a dosimeter, the doctor can even monitor the efficacy of prescribed medications, and the physiological response to medications, over time. This dosimetry approach is directly applicable to clinical studies of various treatments. For example, during a clinical trial, the earpiece module 201 can collect environmental data, drug dosimetry data, and physiological data from the earpiece user such that researchers can understand the etymology between drugs, genes, physiology, environment, and personal health.
Because of the high compliance of the earpiece module 100, primarily due to the dual-modality as a health/environmental monitor and a personal communication/entertainment device, users are prone to wear this device throughout clinical trials, providing more valuable information for drug discovery and the pharmaceuticals market.
As a further example, the health and environmental network 200 can be used by dieticians to track the caloric intake, health, and physical activity of dieters. Similarly, the network 200 can be used by athletic trainers to monitor the diet, physical activity, health, and environment of athletes. In many cases professionals are not necessary, and the user can monitor his/her own diet, activity, athletic performance, etc. through the network without professionals, parents, guardians, or friends monitoring their personal statistics.
In a specific example of the network 200, an earpiece user is a test subject in a clinical trial for a new treatment, such as a new drug, physical therapy, medical device, or the like. The earpiece user's health and environment are monitored in real-time, and this data is stored on the earpiece module 201, the portable telecommunication device 210, the personal database 212, or the anonymous database 213. By accessing the stored data, researchers managing the clinical trial can then compare the statistics from multiple users to make correlations between user environment, health, and the effectiveness of treatment.
It should be noted that algorithms for processing personal health and environmental data, diagnosing medical conditions, assessing health states, and the like do not need to be limited to the illustrated network 200. Various algorithms can also be integrated into the earpiece module 201 or telecommunication device 210 according to embodiments of the present invention. A data storage component in at least one of these units allows processed signal data to be stored, analyzed, and manipulated to provide new knowledge to the user. This storage component can be any solid-state storage device, such as flash memory, random-access memory (RAM), magnetic storage, or the like. For example, the earpiece module 201 can be programmed to monitor certain habits, such as nail-biting. In this non-limiting example, the earpiece module physiological sensors 101 may monitor internal sounds, and an algorithm can be implemented to monitor signatures of nail-biting sounds in real-time. If the habit is identified by the algorithm, the earpiece communication module 107 may instantly warn the user that the habit is occurring. Alternatively, the algorithm may count the number of times a day the habit occurred, monitor physiological and psychological stress indicators during each occurrence, log each time when the habit occurred, and store environmental data associated with the habit. This stored data can be accessed at a later time, allowing the user to determine what environmental factors cause the physiological or psychological stress associated with nail-biting. As this example shows, these algorithms can take advantage of both physiological sensor 101 data and environmental sensor 102 data.
A data storage component may include various algorithms, without limitation. In some embodiments, at least one algorithm is configured to focus processing resources on the extraction of physiological and/or environmental information from the various environmental and/or physiological sensors. Algorithms may be modified and/or uploaded wirelessly via a transmitter (e.g., receiver/transmitter 104 of the earpiece module 100)
The biofeedback functionality of the telemetric earpiece module 100 can be applied towards various gaming applications. For example, one or more earpiece users can connect their earpiece modules 100 to one or more gaming devices wirelessly through the open architecture network provided by Bluetooth, ZigBee, or other such networks. This allows personal health and environmental information to be transferred wirelessly between the earpiece module 100 and a gaming device. As earpiece users play a game, various personal health and environmental feedback can be an active component of the game. In a non-limiting embodiment, two users playing a dancing game, such as Dance Dance Revolution, can monitor their vital signs while competing in a dancing competition. In some cases, users having healthier vital signs, showing improved athletic performance, will get extra points (“Vital Points”). In another specific example, this personal health and environmental information can be sent telemetrically to a gaming device to make entertaining predictions about one or more users. Namely, the gaming device may predict someone's life expectancy, love-life, future occupation, capacity for wealth, and the like. These predictions can be true predictions, purely entertaining predictions, or a mixture of both. Sensors measuring external stressors (such as outside noise, lighting conditions, ozone levels, etc.) and sensors measuring internal stresses (such as muscle tension, breathing rate, pulse rate, etc.) integrated into the earpiece module 100 can be used to facilitate predictions by the gaming device. For example, the information from the sensors can be recorded from one or more earpiece users during a series of questions or tasks, and the information can be sent telemetrically to a gaming device. An algorithm processed in the gaming device can then generate an entertaining assessment from the information. This game can be in the form of a video game, with a graphical user interface 214, or it can be a game “in person” through an entertainer. Other games can involve competitions between multiple earpiece monitor users for health-related purposes, such as online dieting competitions, fitness competitions, activity competitions, or the like. Combining the telemetric earpiece module 100 with gaming, according to embodiments of the present invention, provides seamless interaction between health and environmental monitoring and the game, through a comfortable telemetric module. Other sensor modules 201 located at other parts of the body can also be used.
An additional non-limiting embodiment of the biofeedback functionality of the earpiece module 201 can be monitoring psychological and physiological stress (such as monitoring stress indicators) during a poker game. These stress indicators can be breathing rate, muscle tension, neurological activity, brain wave intensity and activity, core body temperature, pulse rate, blood pressure, galvanometric response, and the like. Users may, for example, use the earpiece module 201 to record or display their psychological and physiological stress during a poker game in real-time. This information can be stored or displayed on a portable telecommunication device 210 or sent wirelessly to other parts of the network 200. The user can use this biofeedback to adjust their psychological and physiological stress (or stress indicators) through force of will. This biofeedback process allows earpiece users to self-train themselves to project a certain “poker face,” such as a stoic cold look, a calm cool look, or another preferred look. Additionally, external stressors, such as light intensity and color, external sounds, and ambient temperature, can be sensed, digitized, and transmitted by the earpiece module 100 to a telecommunication device (for storage), providing the user with important information about how the external environment may be affecting their stress response and, hence, poker game. In some games, the stress indicators may be displayed for outside viewers (who are not part of the poker game) as a form of entertainment when watching a group of poker players (each having earpiece modules 201) in a casino, television, or through the Internet.
The biofeedback approach is also directly relevant to personal education as a learning tool. For example, monitoring the physiological and psychological response to learning can be used to help users understand if they are learning efficiently. For example, in the course of reading, the earpiece module 201 can monitor alertness through galvanometric, brainwave, or vital sign monitoring. The user can then use this information to understand what reading methods or materials are stimulating and which are not stimulating to the earpiece user.
In the broader sense, the discussed earpiece-enabled biofeedback method can be used as a self-training tool for improving performance in public speaking, athletic activity, teaching, and other personal and job-related activities.
The health and environmental network 200 enables a variety of additional business methods. For example, users can be charged a fee for downloading or viewing data from the personal and/or anonymous databases 212, 213. Alternatively, users may be allowed free access but be required to register online, providing personal information with no restrictions on use, for the right to view information from the databases. In turn, this personal information can be traded or sold by the database owner(s). This information can provide valuable marketing information for various companies and government interests. The health and environmental data from the databases 212, 213 can be of great value itself, and this data can be traded or sold to others, such as marketing groups, manufacturers, service providers, government organizations, and the like. The web-page or web-pages associated with the personal and anonymous databases 212, 213 may be subject to targeted advertising. For example, if a user shows a pattern of high blood pressure on a personal database 212, a company may target blood pressure treatment advertisements on the user interface 214 (i.e. web page) while the user is logged-in to the personal database through the user interface 214. For example, because various health and environmental statistics of earpiece users will be available on the database, this information can be used to provide a targeted advertising platform for various manufacturers. In this case, a database manager can sell information to others for targeted advertising linked to a user's personal statistics. In some cases, a database owner does not need to sell the statistics in order to sell the targeted advertising medium. As a specific example, a company can provide a database owner with statistics of interest for targeted advertising. For example, the company may request advertising a weight-loss drug to anonymous users having a poor diet, high caloric intake, and/or increasing weight. A database manager can then charge the company a fee for distributing these advertisements to the targeted users as they are logged-in to the database website(s). In this way, the users remain anonymous to the company. Because targeted advertisements can be such a lucrative market, income from these sources may eliminate the need for charging users a fee for logging-in to the databases 212, 213.
The earpiece module 201 and earpiece module network 200 can enable a variety of research techniques. For example, a plurality of earpiece modules 100 worn by users can be used in marketing research to study the physiological and psychological response of test subjects to various marketing techniques. This technique solves a major problem in marketing research: deciphering objective responses in the midst of human subjectivity. This is because the physiological and psychological response of the earpiece user largely represents objective, unfiltered information. For example, users that are entertained by a pilot TV program would have difficulty hiding innate vital signs in response to the program. The data generated by the earpiece module 201 during market research can be transmitted through any component of the telemetric network 200 and used by marketing researchers to improve a product, service, or method.
Another business method provided by the network 200 is to charge users of the network for usage and service (such as compilation service). For example, a clinical trial company may pay a fee for accessing the databases 212, 213 of their test subjects during medical research. In this case, these companies may buy earpiece modules 201 and pay for the service, or the earpiece modules 201 may be provided free to these companies, as the database service can provide a suitable income itself. Similarly, doctors may pay for this service to monitor patients, fire fighters and first responders may pay for this service to monitor personnel in hazardous environments, and athletic trainers may pay for this service to monitor athletes. Also, users can pay for the database service directly themselves. Because these databases 212, 213 are dynamic, updated regularly via the earpiece module 201 of each user, and changing with time for individual users and users en mass, these databases can maintain a long-term value. In other words, there may always be new information on the databases 212, 213.
Another embodiment of the present invention involves methods of combining information from various earpiece health sensors into a meaningful real-time personal health and environmental exposure assessment in a recording device. The meaningful assessment is generated by algorithms that can be executed in the earpiece 201, in the portable telecommunication device 210, or through various other electronic devices and media within the network 200. In one embodiment, raw or preprocessed data from the sensor module 201 is transmitted wirelessly to the telecommunication device 210, and this device executes various algorithms to convert the raw sensor data (from one or more sensors) into a meaningful assessment for the user. In another embodiment these algorithms are executed within the earpiece 201 itself, without the need for processing in the telecommunication device 210. The output from these algorithms can be viewed as charts, graphs, figures, photos, or other formats for the user to view and analyze. Preferably, these formats display various health factors over time with respect to a particular environment, with health factor intensity on the dependent axis and time or environmental factor intensity on the independent axis. However, virtually any relationship between the physiological data and environmental data can be processed by the algorithm, and these relationships can be quantitative, qualitative, or a combination of both.
One innovation involves applying the earpiece sensor module 201 towards a physical or mental health assessment method. An algorithm may combine data from health and environmental sensors 101, 102 towards generating a personal overall health assessment for the user, conditional to a particular environment. For example breathing rate, pulse rate, and core body temperature can be compared with ozone density in the air for generating an ozone-dependent personal health assessment. In another specific example of this innovation, information from the earpiece sensors 101, 102 can be used to monitor overall “mood” of a user in a particular environment. More particularly, algorithmic processing and analyzing of data from sensors for core body temperature, heart rate, physical activity, and lighting condition can provide a personal assessment of overal mood conditional on external lighting conditions.
As previously mentioned, the ear is located at an ideal physiological position for monitoring a variety of health and environmental factors. Thus, the ear location can enable a variety of methodologies for physiological and environmental monitoring with an earpiece module 100. In particular, because the ear canal is naturally designed for the transmission of audible sound, the ear canal facilitates methods for monitoring physiological processes by monitoring internal sounds. However, when extracting physiological information from the body, in a given external environment, environmental information is inevitably part of the extracted signal. This is because external energy, such as external audible noise, is entering the body. Thus, when listening to internal sounds, external sounds are also picked up. A methodology for cleaning up the signal such that it contains clearer information about physiology (as opposed to external environment) is provided by some embodiments of the present invention.
As with processor 103 of
An embodiment of the physiological signal extraction methodology 400 is presented in
The acoustical sensors 501, 502 can contain any acoustical transducer, such as a microphone, piezoelectric crystal, vibrating membrane, magnetic recorder, and the like. Further, the acoustical sensors 501, 502 may contain a variety of layers for filtering sounds and promoting the directional extraction of sound. Additionally, various electrical filters, such as low-pass, high-pass, band-pass, notch, and other filters, can be used to clean-up signals from each sensor 501, 502 to help remove unwanted sounds or signatures. In some embodiments, the acoustical sensors 501, 502 are compact microphones, such as electric microphones or piezoelectric microphones, and the signals from these microphones are electrically filtered. The acoustical buffer region 519 can be any material that absorbs acoustical energy. In some embodiments, this material is soft, durable material, such as plastic, foam, polymer, or the like. In some embodiments, the acoustical buffer region 519 can be specially shaped to reflect or absorb sounds of certain frequencies through acoustical interference.
An example of how the auscultatory signal extraction technique 500 may be used is summarized in test data 800 set forth in
Another embodiment of the signal extraction methodology 400 is presented in
Because the thin tympanic membrane 620 vibrates significantly in response to sound, whereas the other physiological ear features, such as the ear canal 630 and external ear 606 do not vibrate significantly, a method of extracting secondary optical signals scattered from the vibrating tympanic membrane 620 is provided. Acoustical information from the tympanic membrane vibrational response, collected by the acoustical sensor 601 (or a combination of sensors 601, 602), is processed by a signal processor 405, and the secondary optical information from the tympanic membrane is collected by the optical detector(s) 625. The signal processor compares digitized signals from these sensors in synchronized time, such that signals from the optical detector 625 containing frequency components characteristic of the tympanic membrane's vibrational response are selectively extracted to provide cleaner physiological information from the tympanic membrane. For example, the oxygen content of blood in the tympanic membrane can be monitored by the reflection of red and infrared light from the tympanic membrane, similar to pulse-oximetry. However, scattered optical energy from the ear canal may make it difficult to extract blood oxygen from the tympanic membrane, as the source of scattered light is unclear. The optical physiological signal extraction methodology 600 provides a means of locking-in to the tympanic membrane optical reflection signal through the tympanic membrane vibrational signal collected by the acoustical sensor 601 (or the combination of sensors 601, 602). The illustrated methodology 600 works because the thin tympanic membrane 620, in response to sound, vibrates with substantially greater deflection than the ear canal, and thus primary and secondary light scattered from the tympanic membrane 620 is largely modulated by the frequency of sound reaching the tympanic membrane. This same technique 600 can be applied towards monitoring scattered optical energy from the vibrating bones of the ear, using electromagnetic wavelengths capable of passing through the tympanic membrane.
Another non-limiting embodiment of the optical physiological signal extraction methodology 600 of the present invention involves dosimetry. For example, the concentration of a drug and/or the performance of a drug can be monitored over time by characterizing the real-time fluorescence response of the drug, or intentional fluophores placed in the drug, via the tympanic membrane. In such case, the fluorescence response of the tympanic membrane can be extracted from optical noise through the illustrated methodology 600. However, in this case the fluorescence response, as opposed to the reflectance response, is extracted from the tympanic membrane 620.
In the optical physiological signal extraction methodology 600, a variety of devices can be used for the optical source or optical sources 624, such as a light-emitting diode (LED), a laser diode (LD), a miniature lamp (such as a miniature incandescent lamp, a miniature mercury lamp, or a miniature tungsten lamp), a light guide delivering light from an outside source (such as the sun or other light source), a multi-wavelength source, a microplasma source, an arc source, a combination of these sources, and the like. Special variants of light-emitting diodes, such as resonant-cavity light emitting diodes (RCLEDs), superluminescent LEDs (SLEDs), organic LEDs (OLEDs), and the like can also be applied.
In some embodiments of
It should be noted that in some cases the optical physiological signal extraction methodology 600 can be implemented without the optical emitter 624. For example, the native IR blackbody radiation of the tympanic membrane, scattered in proportion to acoustical vibrational motion of the tympanic membrane, can be extracted using the approach 600 of
A specific pictorial example of the innovative optical physiological signal extraction methodology 600, incorporating an LED-photodetector module 700, is shown in
An optical reflector 727 may be used to steer the light from the LED 724 towards the tympanic membrane 720 and away from the photodetector 725, preventing convolution and saturation by the optical source light 713. The LED 724 and photodetector 725 are mounted onto a mounting board 726 in a discrete module, and this discrete module may be mounted to a larger board 710 for integration with circuitry in an earpiece sensor module 100. Mounting of components to the board 726 and the board 726 to the larger board 710 can be accomplished, for example, by heating soldering bumps underneath the parts through standard electronic soldering techniques. The photodetector 724 can be any solid state device, such as a photodiode, an avalanche photodiode, a photoconductor, a photovoltaic, a photomultiplier, a FET photodetector, a photomultiplier tube, or the like. In some cases, the larger mounting board 710 may be connected to a detachable element, such as a cable, jack, fixture, or the like.
The active optical absorption region of the photodetector 725 may be covered by at least one optical filter for selectively passing light of interest. Light-guiding optics may also be integrated. Optical filters and light-guiding optics may also be applied to the LED source 724. The LED 724 can be any optical wavelength from the deep-UV to the deep-IR. In some cases, the LED 724 can be replaced with a laser diode or other compact laser source, as long as electrical powering requirements are satisfied. In such case, the laser diode may need to be pulsed on a set interval to prevent a battery drain from continuous laser diode usage.
Referring back to
According to some embodiments of the present invention, a person's vitals signs can be extracted through the same methodology, but in this case, the sounds measured from at least one external energy sensor (Sensor-2, Sensor-3, or Sensor-4 404) are also deconvoluted from the final signal such that the final signal contains cleaner physiological information than that from Sensor-1 401 alone. As a further example of the acoustical signal extraction methodology 500 of the present invention, the signal extraction technique can be used to extract acoustical signals associated with one or more of the following: yawning, swallowing, eating, masticating, sleeping, slurping, walking, running, physical activity, jogging, jumping, teeth grinding, jaw movements, a change in bite, changes in speech, changes in voice (volume, pitch, speed, inflammation of vocal chords, etc.), coughing, snoring, sneezing, laughing, eye muscle movements, crying, yelling, vocal stress, physical and psychological stress, stuttering, digestion, organ functioning, vital signs, pulse rate, breathing rate, cardiovascular performance, pulmonary performance, lung capacity, breathing volume, blood pressure, athletic performance, physiological or psychological stress indicators, the number of typed words on a keyboard or typing rate, personal habits (such as scratching, nail biting, saying “um,” hair pulling, smoking, and the like), emotional states, muscle tension, and the like.
It should be clear that the general physiological signal extraction methodology 500 is also applicable in the reverse. Namely, the external environmental energy can be extracted from the convolution of external energy with physiological energy through the same basic process. In such case, the signal processor 405 subtracts signatures associated with internal physiological energy such that the new processed signal contains cleaner information about the environment. It should also be clear that any of the sensors 401, 402, 403, and 404 can be composed of multiple sensors measuring multiple forms and expressions of various physical energies.
The earpiece modules described herein need not be embodied within headsets. For example, an earpiece module 100 according to embodiments of the present invention can be a hearing aid, an earplug, an entertaining speaker, the earpiece for an IPOD, Walkman, or other entertainment unit, a commercial headset for a phone operator, an earring, a gaming interface, or the like. The earpiece module 100 covers the broad realm of earpieces, ear jewelry, and ear apparatuses used by persons for entertainment, hearing, or other purposes both inside and outside of health and environmental monitoring.
Moreover, two earpiece modules 100 may be utilized, one for each ear of a person, according to some embodiments of the present invention. Dual-ear analysis with two earpiece modules can be used, for example, to compare the core temperature of each tympanic membrane in order to gauge brain activity comparing each brain hemisphere. In another case, acoustical energy, including ultrasonic energy, can be passed from one earpiece module to the other, with acoustic absorption and reflection being used to gauge various physiological states. For example, this technique can be used to gauge hydration level in the head or brain by estimating the acoustical energy absorption rate and sound velocity through the head of the user.
A variety of earpiece styles, shapes, and architectures can be used for earpiece module 100 according to embodiments of the present invention. A non-limiting embodiment of the earpiece module is shown pictorially in
It should be understood that the earpiece body 902 can be any shape and size suitable for supporting an earpiece fitting 1008. In some cases, the earpiece body and earpiece fitting can be one and the same structure, such that the earpiece body-fitting is a small fitting inside the ear. In many cases, it is desirable to seal off or partially seal off the ear canal so as to prevent sounds from entering or leaving the ear such that auscultatory signal can more easily be extracted from the ear canal through devices (such as microphones) in the earpiece body-fitting.
It should be noted that the invention is not limited to the exemplary earpiece 905 of
Galvanometry, the measurement of electrical properties of the skin, can be measured inductively, through contactless electromagnetic induction without contacts, or conductively, with two, three, four, or more conductivity probes. Additionally, a 4-point conductivity probe technique, such as that used for measuring the conductivity of semiconductor wafers, can be applied. A variety of sensors can be integrated into the earpiece fitting 1208. For example, a galvanometric device can be integrated into the surface 1209 of the earpiece fitting where the earpiece fitting touches the skin of the outer ear. Additionally, an inductive device, such as an inductive coil 1214, can be integrated along the earpiece fitting body to measure movements of the tympanic membrane inductively. The inductive impedance can also be measured with the inductive coil 1214 or another inductive sensor, and this can be applied towards contactless galvanometry. The inductive coil 1214 can include one or more coils arranged in any orientation, and a core material, such as an iron-containing material, may be used to improve the sensitivity of the response. In some cases, multiple coils may be used to facilitate the canceling of stray electromagnetic interference. Sensors can also be integrated into the end tip 1212 of the earpiece fitting 1208 to measure physiological properties deeper into the ear canal. For example, the optical module 700 of
Alignment, placement, and arrangement of sensors, according to embodiments of the present invention, can be enabled or simplified by adopting a flexible circuitry configuration 1300, such as that shown in
Another multifunctional earpiece module 1500, according to embodiments of the present invention, is illustrated in
The person illustrated in
Embodiments of the present invention are not limited to earpiece modules. Other types of modules may be utilized that attach to other portions of a person's body. For example, a temple module 1600 having a similar design as the earpiece module design 100 can also be employed, as illustrated in
Pulse oximetry is a standard noninvasive technique of estimating blood gas levels. Pulse oximeters typically employ 2 or more optical wavelengths to estimate the ratio of oxygenated to deoxygenated blood. Similarly, various types of hemoglobin, such as methemoglobin and carboxyhemoglobin can be differentiated by measuring and comparing the optical absorption at key red and near-infrared wavelengths.
Though only two optical wavelengths are shown emanating from the source 1708, several additional wavelengths can be incorporated and/or replace conventional wavelengths. For example, by adding additional visible and infrared wavelengths, myoglobin, methemoglobin, carboxyhemoglobin, bilirubin, SpCO2, and blood urea nitrogen (BUN) can be estimated and/or monitored in real-time in addition to the conventional pulse oximetry SpO2 measurement.
The optical source 1708 can include light-emitting diodes (LEDs), laser diodes (LDs), or other compact optical sources. In some cases, optical energies from the optical sources can be guided with waveguides, such as fiber optics. In some cases, ambient light, such as room light or solar radiation, may be sufficient for the optical source 1708. In such case, waveguides may be used to couple ambient light towards the earlobe or other point of interest, Ambient light may be useful in that ambient light may represent a diffuse optical source that is largely independent of body position, such that motion artifacts associated with body motion may be lessened. The optical detectors 1709 can include photodiodes (PDs), avalanche photodiodes (APDs), photomultipliers, or other compact optical detectors.
The reflective pulse oximetry setup 1817 is also suitable for measuring fluorescence from various skin or blood analytes. For example, the optical sources and/or photodetectors may be selectively filtered to measure key fluorescence bands. A fluorescence approach can be applied to, for example, the real-time monitoring of cholesterol and lipids in the skin or blood. Though the optical techniques of
Blood hydration can also be monitored optically, as water selectively absorbs optical wavelengths in the mid-IR and blue-UV ranges, whereas water can be more transparent to the blue-green wavelengths. Thus, the same optical emitter/detector configuration used in earpiece pulse oximetry (
A variety of techniques can be used for monitoring blood metabolites via an earpiece module, such as earpiece module 100. For example, glucose can be monitored via iontophoresis at the surface of the skin combined with enzyme detection. Blood urea nitrogen (BUN) can be monitored by monitoring UV fluorescence in blood (through the skin) or by monitoring visible and mid-IR light absorption using the pulse oximetry approach described above. Various ions such as sodium, potassium, magnesium, calcium, iron, copper, nickel, and other metal ions, can be monitored via selective electrodes in an earpiece module following iontophoresis through the skin. The optical physiological signal extraction approach 600 described above can be used to monitor glucose from the tympanic membrane by monitoring optical reflection and optical fluorescence from the tympanic membrane in response to IR and blue light.
Cardiopulmonary functioning can be evaluated by monitoring blood pressure, pulse, cardiac output, and blood gas levels via earpiece modules, and other monitoring apparatus in accordance with some embodiments of the present invention. Pulse rate and intensity can be monitored through pulse oximetry (described above) as well as by sensing an increase in oxygenated blood with time. Pulse rate and blood flow may also be assessed through impedance measurements via galvanometry near a blood vessel. Additionally, pulse rate and blood flow may be assessed through a fast-response thermal energy sensor, such as a pyroelectric sensor. Because moving blood may temporarily increase or decrease the localized temperature near a blood vessel, a pyroelectric sensor will generate an electrical signal that is proportional to the total blood flow in time. Blood pressure can be monitored along the earlobe, for example. According to some embodiments of the present invention, a digital blood pressure meter is integrated into an earpiece module, such as earpiece 905 of
Organ function monitoring includes monitoring, for example, the liver, kidneys, pancreas, skin, and other vital or important organs. Liver quality can be monitored noninvasively by monitoring optical absorption and reflection at various optical wavelengths. For example, optical reflection from white LEDs or selected visible-wavelength LEDs can be used to monitor bilirubin levels in the skin and blood, for a real-time assessment of liver health.
Monitoring neurological functioning can be accomplished via electrodes placed at the ear, near the ear, or along another surface of the body. When such electrodes are placed along the forehead, this process is described as electroencephalography, and the resulting data is called an electroencephalogram (EEG). These electrodes can be either integrated into an earpiece module or connected to an earpiece module, according to some embodiments of the present invention. For example, an earlobe clip (e.g., 904,
A person's body motion and head position can be monitored by integrating a motion sensor into an earpiece module (e.g., 905,
Body temperature, including core and skin temperature, can be monitored in real-time by integrating compact infrared sensors into an earpiece module, according to some embodiments of the present invention. Infrared sensors are generally composed of thermoelectric/pyroelectric materials or semiconductor devices, such as photodiodes or photoconductors. Thermistors, thermocouples, and other temperature-dependent transducers can also be incorporated for monitoring body temperature. These sensors can be very compact and thus can be integrated throughout an earpiece module. In some embodiments, these sensors may be mounted along the backside of an earpiece body, as illustrated in
In some embodiments of the present invention, a pedometer can be integrated into an earpiece module to measure the number of steps walked during a day. Pedometers that can be integrated into an earpiece module include, but are not limited to, mechanical pedometers (usually implementing a metallic ball or spring), microelectromechanical systems (MEMS) pedometers, inertial sensor pedometers, accelerometer-based pedometers, accelerometry, gyroscopic pedometers, and the like.
In some embodiments of the present invention, a pedometer for an earpiece module employs an acoustic sensor for monitoring the characteristic sounds of footsteps channeled along the ear canal. For example, an acoustic sensor can be integrated into an earpiece housing (e.g., 902,
Breathing characteristics can be monitored in a manner similar to that of acoustic pedometry (described above) in the auscultatory extraction methodology 500. In some embodiments, an acoustic sensor in an earpiece module is used to sense sounds associated with breathing. Signal processing algorithms are then used to extract breathing sounds from other sounds and noise. This information is processed into a breathing monitor, capable of monitoring, for example, the intensity, volume, and speed of breathing. Another method of monitoring breathing is to employ pressure transducers into an earpiece module. Changes in pressure inside or near the ear associated with breathing can be measured directly and, through signal processing, translated into a breathing monitor. Similarly, optical reflection sensors can be used to monitor pressure in or near the ear by monitoring physical changes in the skin or tissues in response to breathing. For monitoring the physical changes of the tympanic membrane in response to breathing, and hence ascertaining breathing rate, the optical signal extraction approach 600 described above can be employed. At least one color sensor, or colormetric sensor, can be employed to monitor changes in color associated with breathing and other health factors. In the various embodiments described herein, the location of these acoustic sensors is in or near an earpiece fitting (e.g., 1008,
It should be noted that some embodiments of the present invention incorporate health sensors that do not employ chemical or biological reagents for monitoring various health factors. This is because such sensors have traditionally required larger instrumentation (not suitable for portability) and/or disposable samplers (not acceptable to most end users). However, sensors employing chemical or biological reagents may be incorporated into earpiece modules, according to some embodiments of the present invention. For example, the diffusion of analyte through the skin can be monitored electrically or optically by selective binding to enzymes or antibodies contained in the health sensors integrated into an earpiece module. In some cases, iontophoresis, agitation, heat, or osmosis may be required to pull ions from the skin or blood into the sensor region for monitoring health factors. In some cases, these analytes may be tagged with markers for electromagnetic, electrical, nuclear, or magnetic detection.
Caloric intake, physical activity, and metabolism can be monitored using a core temperature sensor, an accelerometer, a sound extraction methodology (e.g., 500,
Environmental temperature can be monitored, for example, by thermistor, thermocouple, diode junction drop reference, or the like. Electrical temperature measurement techniques are well known to those skilled in the art, and are of suitable size and power consumption that they can be integrated into a wireless earpiece module without significant impact on the size or functionality of the wireless earpiece module.
Environmental noise can be monitored, for example, by transducer, microphone, or the like. Monitoring of environmental noise preferably includes, but is not limited to, instantaneous intensity, spectral frequency, repetition frequency, peak intensity, commonly in units of decibels, and cumulative noise level exposures, commonly in units of decibel-hours. This environmental noise may or may not include noise generated by a person wearing an earpiece module. Sound made by a person wearing an earpiece module may be filtered out, for example, using analog or digital noise cancellation techniques, by directional microphone head shaping, or the like. The environmental noise sensor may or may not be the same sensor as that used for the intended purpose of wireless communication. In some embodiments, the environmental noise sensor is a separate sensor having broader audible detection range of noise level and frequency, at the possible sacrifice of audio quality.
Environmental smog includes VOC's, formaldehyde, alkenes, nitric oxide, PAH's, sulfur dioxide, carbon monoxide, olefins, aromatic compounds, xylene compounds, and the like. Monitoring of the aforementioned smog components can be performed using earpiece modules and other wearable apparatus, according to embodiments of the present invention, in a variety of methods. All smog components may be monitored. Alternatively, single smog components or combinations of smog components may be monitored. Photoionization detectors (PID's) may be used to provide continuous monitoring and instantaneous readings. Other methods of detecting smog components according to embodiments of the present invention include, but are not limited to, electrocatalytic, photocatalytic, photoelectrocatalytic, colorimetric, spectroscopic or chemical reaction methods. Examples of monitoring techniques using the aforementioned methods may include, but are not limited to, IR laser absorption spectroscopy, difference frequency generation laser spectroscopy, porous silicon optical microcavities, surface plasmon resonance, absorptive polymers, absorptive dielectrics, and colorimetric sensors. For example, absorptive polymer capacitors inductors, or other absorptive polymer-based electronics can be incorporated into an earpiece module (e.g., 100,
In some embodiments of the present invention, environmental air particles can be monitored with a flow cell and a particle counter, particle sizer, particle identifier, or other particulate matter sensor incorporated as part of an earpiece module (e.g., 100,
In some embodiments of the present invention, environmental air pressure or barometric pressure can be monitored by a barometer. Non-limiting examples of barometric pressure measurement include hydrostatic columns using mercury, water, or the like, foil-based or semiconductor-based strain gauge, pressure transducers, or the like. In some embodiments of the present invention, semiconductor-based strain gauges are utilized. A strain gauge may utilize a piezoresistive material that gives an electrical response that is indicative of the amount of deflection or strain due to atmospheric pressure. Atmospheric pressure shows a diurnal cycle caused by global atmospheric tides. Environmental atmospheric pressure is of interest for prediction of weather and climate changes. Environmental pressure may also be used in conjunction with other sensing elements, such as temperature and humidity to calculate other environmental factors, such as dew point. Air pressure can also be measured by a compact MEMS device composed of a microscale diaphragm, where the diaphragm is displaced under differential pressure and this strain is monitored by the piezoelectric or piezoresistive effect.
In some embodiments of the present invention, environmental humidity, relative humidity, and dew point can be monitored by measuring capacitance, resistivity or thermal conductivity of materials exposed to the air, or by spectroscopy changes in the air itself. Resistive humidity sensors measure the change in electrical impedance of a hygroscopic medium such as a conductive polymer, salt, or treated substrate. Capacitive humidity sensors utilize incremental change in the dielectric constant of a dielectric, which is nearly directly proportional to the relative humidity of the surrounding environment. Thermal humidity sensors measure the absolute humidity by quantifying the difference between the thermal conductivity of dry air and that of air containing water vapor. Humidity data can be stored along with pressure monitor data, and a simple algorithm can be used to extrapolate the dew point. In some embodiments of the present invention, monitoring humidity is performed via spectroscopy. The absorption of light by water molecules in air is well known to those skilled in the art. The amount of absorption at known wavelengths is indicative of the humidity or relative humidity. Humidity may be monitored with a spectroscopic method that is compatible with the smog monitoring spectroscopic method described above.
When environmental factors such as the aforementioned are monitored continuously in real-time, a user's total exposure level to an environmental factor can be recorded. When a representative volume of air a user has been exposed to is monitored or estimated, the volumetric concentration of the analytes can be calculated or estimated. In order to estimate the volume of air a person wearing the earpiece has been exposed to, a pedometer or accelerometer or air flow sensor can also be integrated into an earpiece module. Pedometers and accelerometers can be integrated into an earpiece module via mechanical sensors (usually implementing a mechanical-electrical switch), MEMS devices, and/or gyroscopic technologies. The technologies required for these types of pedometers and accelerators are well known to those skilled in the art. The incorporated pedometer or accelerometer (or more than one pedometer or accelerometer) is used to gage the distance a person has traveled, for use in the estimation of the volume of air to which a person has been exposed, and the subsequent estimate of the volumetric concentration of monitored analytes.
The health and environmental sensors utilized with earpiece modules and other wearable monitoring apparatus, according to embodiments of the present invention, can operate through a user-selectable switch on an earpiece module. However, health and environmental sensors can also be run automatically and independently of the person wearing the apparatus. In other embodiments, the person may control health and environmental monitoring through a device wirelessly coupled to an earpiece module, such as a portable telecommunication device (e.g., 210,
The earpiece module 100 may be configured such that user preferences can be “downloaded” wirelessly without requiring changes to the earpiece monitor hardware. For example, an earpiece concerned about a heart condition may wish to have the signal processor 103 focus on processing pulse signature, at the expense of ignoring other physiological or environmental parameters. The user may then use the portable telecommunication device 210 to download a specialized algorithm through the web. This may be accomplished through existing wireless infrastructure by text-messaging to a database containing the algorithm. The user will then have an earpiece module suited with analysis software specialized to the needs and desires of the user.
Health and environmental monitors, according to embodiments of the present invention, enable low-cost, real-time personal health and environmental exposure assessment monitoring of various health factors. An individual's health and environmental exposure record can be provided throughout the day, week, month, or the like. Moreover, because the health and environmental sensors can be small and compact, the overall size of an apparatus, such as an earpiece, can remain lightweight and compact.
The foregoing is illustrative of the present invention and is not to be construed as limiting thereof. Although a few exemplary embodiments of this invention have been described, those skilled in the art will readily appreciate that many modifications are possible in the exemplary embodiments without materially departing from the teachings and advantages of this invention. Accordingly, all such modifications are intended to be included within the scope of this invention as defined in the claims. The invention is defined by the following claims, with equivalents of the claims to be included therein.
This application is a continuation application of U.S. patent application Ser. No. 14/063,669, filed Oct. 25, 2013, which is a continuation application of U.S. patent application Ser. No. 11/811,844, filed Jun. 12, 2007, now U.S. Pat. No. 8,652,040, and which claims the benefit of and priority to U.S. Provisional Patent Application No. 60/905,761, filed Mar. 8, 2007, U.S. Provisional Patent Application No. 60/876,128, filed Dec. 21, 2006, and U.S. Provisional Patent Application No. 60/875,606, filed Dec. 19, 2006, the disclosures of which are incorporated herein by reference as if set forth in their entireties.
Number | Name | Date | Kind |
---|---|---|---|
3595219 | Friedlander et al. | Jul 1971 | A |
3922488 | Gabr | Nov 1975 | A |
4005701 | Aisenberg et al. | Feb 1977 | A |
4025734 | Aloupis | May 1977 | A |
4240882 | Ang et al. | Dec 1980 | A |
4312358 | Barney | Jan 1982 | A |
4331154 | Broadwater et al. | May 1982 | A |
4438772 | Slavin | Mar 1984 | A |
4459645 | Glatter | Jul 1984 | A |
4491760 | Linvill | Jan 1985 | A |
4521499 | Switzer | Jun 1985 | A |
4541905 | Kuwana et al. | Sep 1985 | A |
4592807 | Switzer | Jun 1986 | A |
4598417 | Deno | Jul 1986 | A |
4655225 | Dahne et al. | Apr 1987 | A |
4736431 | Allie et al. | Apr 1988 | A |
4783815 | Büttner | Nov 1988 | A |
4830014 | Goodman et al. | May 1989 | A |
4850962 | Schaefer | Jul 1989 | A |
4878501 | Shue | Nov 1989 | A |
4882492 | Schlager | Nov 1989 | A |
4896676 | Sasaki | Jan 1990 | A |
4928704 | Hardt | May 1990 | A |
4952890 | Swanson | Aug 1990 | A |
4952928 | Carroll et al. | Aug 1990 | A |
4957109 | Groeger et al. | Sep 1990 | A |
4985925 | Langberg et al. | Jan 1991 | A |
5002060 | Nedivi | Mar 1991 | A |
5022970 | Cook et al. | Jun 1991 | A |
5025791 | Niwa | Jun 1991 | A |
5046103 | Warnaka et al. | Sep 1991 | A |
5079421 | Knudson et al. | Jan 1992 | A |
5080098 | Willett et al. | Jan 1992 | A |
5086229 | Rosenthal et al. | Feb 1992 | A |
5091954 | Sasaki et al. | Feb 1992 | A |
5115133 | Knudson | May 1992 | A |
5119819 | Thomas et al. | Jun 1992 | A |
5131047 | Hashimoto et al. | Jul 1992 | A |
5138663 | Moseley | Aug 1992 | A |
5139025 | Lewis et al. | Aug 1992 | A |
5143078 | Mather et al. | Sep 1992 | A |
5179951 | Knudson | Jan 1993 | A |
5182774 | Bourk | Jan 1993 | A |
5226417 | Swedlow et al. | Jul 1993 | A |
5237994 | Goldberger | Aug 1993 | A |
5259033 | Goodings et al. | Nov 1993 | A |
5299570 | Hatschek | Apr 1994 | A |
5309922 | Schechter et al. | May 1994 | A |
5348002 | Caro | Sep 1994 | A |
5377100 | Pope et al. | Dec 1994 | A |
5402496 | Soli et al. | Mar 1995 | A |
5444786 | Raviv | Aug 1995 | A |
5448082 | Kim | Sep 1995 | A |
5467775 | Callahan et al. | Nov 1995 | A |
5469855 | Pompei et al. | Nov 1995 | A |
5471009 | Oba et al. | Nov 1995 | A |
5481615 | Eatwell et al. | Jan 1996 | A |
5482036 | Diab et al. | Jan 1996 | A |
5492129 | Greenberger | Feb 1996 | A |
5494043 | O'Sullivan et al. | Feb 1996 | A |
5499301 | Sudo et al. | Mar 1996 | A |
5539831 | Harley | Jul 1996 | A |
5572990 | Berlin | Nov 1996 | A |
5581648 | Sahagen | Dec 1996 | A |
5596987 | Chance | Jan 1997 | A |
5652570 | Lepkofker | Jul 1997 | A |
5662117 | Bittman | Sep 1997 | A |
5671301 | Kupershmidt | Sep 1997 | A |
5673692 | Schulze et al. | Oct 1997 | A |
5697374 | Odagiri et al. | Dec 1997 | A |
5704365 | Albrecht et al. | Jan 1998 | A |
5711308 | Singer | Jan 1998 | A |
5721783 | Anderson | Feb 1998 | A |
5722418 | Bro | Mar 1998 | A |
5725480 | Oosta et al. | Mar 1998 | A |
5743260 | Chung et al. | Apr 1998 | A |
5779631 | Chance | Jul 1998 | A |
5797841 | Delonzor et al. | Aug 1998 | A |
5807114 | Hodges et al. | Sep 1998 | A |
5807267 | Bryars et al. | Sep 1998 | A |
5817008 | Rafert et al. | Oct 1998 | A |
5846190 | Woehrle | Dec 1998 | A |
5853005 | Scanlon | Dec 1998 | A |
5873836 | Kahn et al. | Feb 1999 | A |
5881159 | Aceti | Mar 1999 | A |
5904654 | Wohltmann et al. | May 1999 | A |
5938593 | Quellette | Aug 1999 | A |
5954644 | Dettling et al. | Sep 1999 | A |
5964701 | Asada et al. | Oct 1999 | A |
5971931 | Raff | Oct 1999 | A |
5974338 | Asano et al. | Oct 1999 | A |
5995858 | Kinast | Nov 1999 | A |
6004274 | Aceti et al. | Dec 1999 | A |
6013007 | Root et al. | Jan 2000 | A |
6022748 | Charych et al. | Feb 2000 | A |
6023541 | Merchant et al. | Feb 2000 | A |
6030342 | Amano et al. | Feb 2000 | A |
6045511 | Ott et al. | Apr 2000 | A |
6052336 | Lowrey, III | Apr 2000 | A |
6067006 | O'Brien | May 2000 | A |
6070093 | Oosta et al. | May 2000 | A |
6078829 | Uchida et al. | Jun 2000 | A |
6080110 | Thorgersen | Jun 2000 | A |
6081742 | Amano et al. | Jun 2000 | A |
6144867 | Walker et al. | Nov 2000 | A |
6155983 | Kosuda et al. | Dec 2000 | A |
6168567 | Pickering et al. | Jan 2001 | B1 |
6186145 | Brown | Feb 2001 | B1 |
6198394 | Jacobsen et al. | Mar 2001 | B1 |
6198951 | Kosuda et al. | Mar 2001 | B1 |
6205354 | Gellermann et al. | Mar 2001 | B1 |
6231519 | Blants et al. | May 2001 | B1 |
6253871 | Aceti | Jul 2001 | B1 |
6267721 | Welles | Jul 2001 | B1 |
6277079 | Avicola et al. | Aug 2001 | B1 |
6283915 | Nolan et al. | Sep 2001 | B1 |
6285816 | Anderson et al. | Sep 2001 | B1 |
6289230 | Chaiken et al. | Sep 2001 | B1 |
6298314 | Blackadar et al. | Oct 2001 | B1 |
6304797 | Shusterman | Oct 2001 | B1 |
6332868 | Sato et al. | Dec 2001 | B1 |
6340350 | Simms | Jan 2002 | B1 |
6358216 | Kraus et al. | Mar 2002 | B1 |
6361660 | Goldstein | Mar 2002 | B1 |
6371925 | Imai et al. | Apr 2002 | B1 |
6373942 | Braund | Apr 2002 | B1 |
6374129 | Chin et al. | Apr 2002 | B1 |
6385176 | Iyengar et al. | May 2002 | B1 |
6409675 | Turcott | Jun 2002 | B1 |
6415167 | Blank et al. | Jul 2002 | B1 |
6443890 | Schulze et al. | Sep 2002 | B1 |
6444474 | Thomas et al. | Sep 2002 | B1 |
6445799 | Taenzer et al. | Sep 2002 | B1 |
6454718 | Clift | Sep 2002 | B1 |
6458080 | Brown et al. | Oct 2002 | B1 |
6470893 | Boesen | Oct 2002 | B1 |
6491644 | Vujanic et al. | Dec 2002 | B1 |
6491647 | Bridger et al. | Dec 2002 | B1 |
6513532 | Mault et al. | Feb 2003 | B2 |
6514278 | Hibst et al. | Feb 2003 | B1 |
6527711 | Stivoric et al. | Mar 2003 | B1 |
6527712 | Brown et al. | Mar 2003 | B1 |
6527729 | Turcott | Mar 2003 | B1 |
6529754 | Kondo | Mar 2003 | B2 |
6534012 | Hazen et al. | Mar 2003 | B1 |
6544199 | Morris | Apr 2003 | B1 |
6556852 | Schulze et al. | Apr 2003 | B1 |
6569094 | Suzuki et al. | May 2003 | B2 |
6571117 | Marbach | May 2003 | B1 |
6605038 | Teller et al. | Aug 2003 | B1 |
6608562 | Kimura et al. | Aug 2003 | B1 |
6616606 | Petersen et al. | Sep 2003 | B1 |
6616613 | Goodman | Sep 2003 | B1 |
6631196 | Taenzer et al. | Oct 2003 | B1 |
6647378 | Kindo | Nov 2003 | B2 |
6656116 | Kim et al. | Dec 2003 | B2 |
6694180 | Boesen | Feb 2004 | B1 |
6702752 | Dekker | Mar 2004 | B2 |
6725072 | Steuer et al. | Apr 2004 | B2 |
6738485 | Boesen | May 2004 | B1 |
6745061 | Hicks et al. | Jun 2004 | B1 |
6748254 | O'Neil et al. | Jun 2004 | B2 |
6760610 | Tschupp et al. | Jul 2004 | B2 |
6783501 | Takahashi et al. | Aug 2004 | B2 |
6808473 | Hisano et al. | Oct 2004 | B2 |
6859658 | Krug | Feb 2005 | B1 |
6893396 | Schulze et al. | May 2005 | B2 |
6941239 | Unuma et al. | Sep 2005 | B2 |
6953435 | Kondo et al. | Oct 2005 | B2 |
6954644 | Johansson et al. | Oct 2005 | B2 |
6996427 | Ali et al. | Feb 2006 | B2 |
6997879 | Turcott | Feb 2006 | B1 |
7011814 | Suddarth | Mar 2006 | B2 |
7018338 | Vetter et al. | Mar 2006 | B2 |
7024369 | Brown et al. | Apr 2006 | B1 |
7030359 | Römhild | Apr 2006 | B2 |
7034694 | Yamaguchi et al. | Apr 2006 | B2 |
7041062 | Friedrichs et al. | May 2006 | B2 |
7043287 | Khalil et al. | May 2006 | B1 |
7054674 | Cane et al. | May 2006 | B2 |
7088234 | Naito et al. | Aug 2006 | B2 |
7088828 | Bradford et al. | Aug 2006 | B1 |
7107088 | Aceti | Sep 2006 | B2 |
7113815 | O'Neil et al. | Sep 2006 | B2 |
7117032 | Childre et al. | Oct 2006 | B2 |
7163512 | Childre et al. | Jan 2007 | B1 |
7175601 | Verjus et al. | Feb 2007 | B2 |
7190986 | Hannula et al. | Mar 2007 | B1 |
7209775 | Bae et al. | Apr 2007 | B2 |
7217224 | Thomas | May 2007 | B2 |
7252639 | Kimura et al. | Aug 2007 | B2 |
7263396 | Chen et al. | Aug 2007 | B2 |
7289837 | Mannheimer et al. | Oct 2007 | B2 |
D555019 | Au Yeung | Nov 2007 | S |
7324668 | Rubinstenn et al. | Jan 2008 | B2 |
7336982 | Yoo et al. | Feb 2008 | B2 |
7341559 | Schultz et al. | Mar 2008 | B2 |
7376451 | Mahony et al. | May 2008 | B2 |
7378954 | Wendt | May 2008 | B2 |
7470234 | Elhag et al. | Dec 2008 | B1 |
7483730 | Diab et al. | Jan 2009 | B2 |
7486988 | Goodall et al. | Feb 2009 | B2 |
7507207 | Sakai et al. | Mar 2009 | B2 |
7519327 | White | Apr 2009 | B2 |
7526327 | Blondeau et al. | Apr 2009 | B2 |
7558625 | Levin et al. | Jul 2009 | B2 |
7583994 | Scholz | Sep 2009 | B2 |
7620450 | Kim et al. | Nov 2009 | B2 |
7625285 | Breving | Dec 2009 | B2 |
7652569 | Kiff et al. | Jan 2010 | B2 |
D611366 | Register et al. | Mar 2010 | S |
7689437 | Teller et al. | Mar 2010 | B1 |
7695440 | Kondo et al. | Apr 2010 | B2 |
D615427 | Au Yeung | May 2010 | S |
7725147 | Li et al. | May 2010 | B2 |
7756559 | Abreu | Jul 2010 | B2 |
7843325 | Otto | Nov 2010 | B2 |
D632984 | Register et al. | Feb 2011 | S |
7881733 | Staton | Feb 2011 | B2 |
7894869 | Hoarau | Feb 2011 | B2 |
7914468 | Shalon et al. | Mar 2011 | B2 |
D635874 | Tseng et al. | Apr 2011 | S |
D637097 | Tseng | May 2011 | S |
7991448 | Edgar et al. | Aug 2011 | B2 |
7998079 | Nagai et al. | Aug 2011 | B2 |
D644542 | Henne et al. | Sep 2011 | S |
8024974 | Bharti et al. | Sep 2011 | B2 |
8050728 | Al-Ali et al. | Nov 2011 | B2 |
8055319 | Oh et al. | Nov 2011 | B2 |
8055330 | Egozi | Nov 2011 | B2 |
8059924 | Letant et al. | Nov 2011 | B1 |
8130105 | Al-Ali et al. | Mar 2012 | B2 |
8137270 | Keenan et al. | Mar 2012 | B2 |
D659569 | Shadovitz | May 2012 | S |
8172459 | Abreu | May 2012 | B2 |
8175670 | Baker, Jr. et al. | May 2012 | B2 |
8204730 | Liu et al. | Jun 2012 | B2 |
8233955 | Al-Ali et al. | Jul 2012 | B2 |
D665679 | Shigeno et al. | Aug 2012 | S |
D665680 | Shigeno et al. | Aug 2012 | S |
8251903 | LeBoeuf et al. | Aug 2012 | B2 |
8255027 | Al-Ali et al. | Aug 2012 | B2 |
8255029 | Addison et al. | Aug 2012 | B2 |
8303512 | Kosuda et al. | Nov 2012 | B2 |
8328420 | Abreu | Dec 2012 | B2 |
D674715 | Dalton | Jan 2013 | S |
8385560 | Solbeck et al. | Feb 2013 | B2 |
D677589 | Jung et al. | Mar 2013 | S |
D677792 | Vandiver | Mar 2013 | S |
8416959 | Lott et al. | Apr 2013 | B2 |
D682126 | Tello | May 2013 | S |
D684071 | Greenwood et al. | Jun 2013 | S |
8491492 | Shinar et al. | Jul 2013 | B2 |
8504679 | Spire et al. | Aug 2013 | B2 |
8506524 | Graskov et al. | Aug 2013 | B2 |
8512242 | LeBoeuf et al. | Aug 2013 | B2 |
D693248 | Anderssen et al. | Nov 2013 | S |
D693249 | Anderssen et al. | Nov 2013 | S |
D693250 | Anderssen et al. | Nov 2013 | S |
D699131 | Marshall et al. | Feb 2014 | S |
D700084 | Hsu | Feb 2014 | S |
8679008 | Hughes et al. | Mar 2014 | B2 |
D702141 | Jung et al. | Apr 2014 | S |
8730048 | Shen et al. | May 2014 | B2 |
9005129 | Venkatraman et al. | Apr 2015 | B2 |
20010000526 | Gopinathan et al. | Apr 2001 | A1 |
20010015123 | Nishitani et al. | Aug 2001 | A1 |
20010027384 | Schulze et al. | Oct 2001 | A1 |
20010039372 | Yasushi et al. | Nov 2001 | A1 |
20010040591 | Abbott et al. | Nov 2001 | A1 |
20010044588 | Mault | Nov 2001 | A1 |
20010049471 | Suzuki et al. | Dec 2001 | A1 |
20010051766 | Gazdzinski | Dec 2001 | A1 |
20020019586 | Teller et al. | Feb 2002 | A1 |
20020021800 | Bodley | Feb 2002 | A1 |
20020035340 | Fraden et al. | Mar 2002 | A1 |
20020099356 | Unger et al. | Jul 2002 | A1 |
20020107649 | Takiguchi et al. | Aug 2002 | A1 |
20020115937 | Song | Aug 2002 | A1 |
20020143242 | Nemirovski | Oct 2002 | A1 |
20020156386 | Dardik et al. | Oct 2002 | A1 |
20020156654 | Roe et al. | Oct 2002 | A1 |
20020165466 | Givens | Nov 2002 | A1 |
20020169485 | Pless et al. | Nov 2002 | A1 |
20020180605 | Ozguz et al. | Dec 2002 | A1 |
20020186137 | Skardon | Dec 2002 | A1 |
20020188210 | Aizawa | Dec 2002 | A1 |
20020194002 | Petrushin | Dec 2002 | A1 |
20030002685 | Werblud | Jan 2003 | A1 |
20030002705 | Boesen | Jan 2003 | A1 |
20030007631 | Bolognesi et al. | Jan 2003 | A1 |
20030036685 | Goodman | Feb 2003 | A1 |
20030045785 | Diab et al. | Mar 2003 | A1 |
20030050563 | Suribhotla et al. | Mar 2003 | A1 |
20030060693 | Monfre | Mar 2003 | A1 |
20030064712 | Gaston et al. | Apr 2003 | A1 |
20030065257 | Mault et al. | Apr 2003 | A1 |
20030065269 | Vetter et al. | Apr 2003 | A1 |
20030083583 | Kovtun et al. | May 2003 | A1 |
20030109030 | Uchida et al. | Jun 2003 | A1 |
20030109791 | Kondo et al. | Jun 2003 | A1 |
20030118197 | Nagayasu et al. | Jun 2003 | A1 |
20030147369 | Singh et al. | Aug 2003 | A1 |
20030147544 | Lichtblau | Aug 2003 | A1 |
20030149526 | Zhou et al. | Aug 2003 | A1 |
20030151524 | Clark | Aug 2003 | A1 |
20030163710 | Ortiz et al. | Aug 2003 | A1 |
20030181795 | Suzuki et al. | Sep 2003 | A1 |
20030181798 | Al-Ali | Sep 2003 | A1 |
20030195040 | Breving | Oct 2003 | A1 |
20030208113 | Mault et al. | Nov 2003 | A1 |
20030212336 | Lee et al. | Nov 2003 | A1 |
20030220584 | Honeyager et al. | Nov 2003 | A1 |
20030222268 | Yocom et al. | Dec 2003 | A1 |
20030233051 | Verjus et al. | Dec 2003 | A1 |
20030234726 | Chen et al. | Dec 2003 | A1 |
20040004547 | Appelt et al. | Jan 2004 | A1 |
20040022700 | Kim et al. | Feb 2004 | A1 |
20040030581 | Leven | Feb 2004 | A1 |
20040032957 | Mansy et al. | Feb 2004 | A1 |
20040034289 | Teller et al. | Feb 2004 | A1 |
20040034293 | Kimball | Feb 2004 | A1 |
20040039254 | Stivoric et al. | Feb 2004 | A1 |
20040054291 | Schulz et al. | Mar 2004 | A1 |
20040075677 | Loyall et al. | Apr 2004 | A1 |
20040077934 | Massad | Apr 2004 | A1 |
20040082842 | Lumba et al. | Apr 2004 | A1 |
20040092846 | Watrous | May 2004 | A1 |
20040103146 | Park | May 2004 | A1 |
20040117204 | Mazar et al. | Jun 2004 | A1 |
20040120844 | Tribelsky et al. | Jun 2004 | A1 |
20040122294 | Hatlestad et al. | Jun 2004 | A1 |
20040122702 | Sabol et al. | Jun 2004 | A1 |
20040133123 | Leonhardt et al. | Jul 2004 | A1 |
20040135571 | Uutela et al. | Jul 2004 | A1 |
20040138578 | Pineda et al. | Jul 2004 | A1 |
20040183675 | Harris | Sep 2004 | A1 |
20040186387 | Kosuda et al. | Sep 2004 | A1 |
20040186390 | Ross et al. | Sep 2004 | A1 |
20040198463 | Knoedgen | Oct 2004 | A1 |
20040203897 | Rogers | Oct 2004 | A1 |
20040212505 | Dewing et al. | Oct 2004 | A1 |
20040215958 | Ellis et al. | Oct 2004 | A1 |
20040219056 | Tribelsky et al. | Nov 2004 | A1 |
20040220483 | Yeo et al. | Nov 2004 | A1 |
20040220488 | Vyshedskiy et al. | Nov 2004 | A1 |
20040225207 | Bae et al. | Nov 2004 | A1 |
20040228494 | Smith | Nov 2004 | A1 |
20040240516 | Harr | Dec 2004 | A1 |
20040242976 | Abreu | Dec 2004 | A1 |
20040254501 | Mault | Dec 2004 | A1 |
20050004458 | Kanayama et al. | Jan 2005 | A1 |
20050007582 | Villers et al. | Jan 2005 | A1 |
20050021519 | Ghouri | Jan 2005 | A1 |
20050027216 | Guillemaud et al. | Feb 2005 | A1 |
20050030540 | Thornton | Feb 2005 | A1 |
20050033200 | Soehren et al. | Feb 2005 | A1 |
20050036212 | Saito | Feb 2005 | A1 |
20050038349 | Choi et al. | Feb 2005 | A1 |
20050043600 | Diab et al. | Feb 2005 | A1 |
20050043630 | Honeyager et al. | Feb 2005 | A1 |
20050058456 | Yoo | Mar 2005 | A1 |
20050059870 | Aceti | Mar 2005 | A1 |
20050084666 | Pong et al. | Apr 2005 | A1 |
20050101845 | Nihtila | May 2005 | A1 |
20050101872 | Sattler et al. | May 2005 | A1 |
20050113167 | Buchner et al. | May 2005 | A1 |
20050113656 | Chance | May 2005 | A1 |
20050113703 | Farringdon et al. | May 2005 | A1 |
20050116820 | Goldreich | Jun 2005 | A1 |
20050119833 | Nanikashvili | Jun 2005 | A1 |
20050134452 | Smith | Jun 2005 | A1 |
20050148883 | Boesen | Jul 2005 | A1 |
20050154264 | Lecompte et al. | Jul 2005 | A1 |
20050163302 | Mock et al. | Jul 2005 | A1 |
20050177029 | Shen | Aug 2005 | A1 |
20050177034 | Beaumont | Aug 2005 | A1 |
20050187448 | Petersen et al. | Aug 2005 | A1 |
20050187453 | Petersen et al. | Aug 2005 | A1 |
20050190072 | Brown et al. | Sep 2005 | A1 |
20050192514 | Kearby et al. | Sep 2005 | A1 |
20050192515 | Givens et al. | Sep 2005 | A1 |
20050192516 | Takiguchi et al. | Sep 2005 | A1 |
20050192557 | Brauker et al. | Sep 2005 | A1 |
20050196009 | Boesen | Sep 2005 | A1 |
20050203349 | Nanikashvili | Sep 2005 | A1 |
20050203357 | Debreczeny et al. | Sep 2005 | A1 |
20050209516 | Fraden | Sep 2005 | A1 |
20050212405 | Negley | Sep 2005 | A1 |
20050222487 | Miller et al. | Oct 2005 | A1 |
20050222903 | Buchheit et al. | Oct 2005 | A1 |
20050226446 | Luo et al. | Oct 2005 | A1 |
20050228244 | Banet | Oct 2005 | A1 |
20050228299 | Banet | Oct 2005 | A1 |
20050228463 | Mac et al. | Oct 2005 | A1 |
20050240087 | Keenan et al. | Oct 2005 | A1 |
20050245839 | Stivoric et al. | Nov 2005 | A1 |
20050258816 | Zen et al. | Nov 2005 | A1 |
20050258950 | Sahashi et al. | Nov 2005 | A1 |
20050259811 | Kimm et al. | Nov 2005 | A1 |
20060004680 | Robarts et al. | Jan 2006 | A1 |
20060009685 | Finarov et al. | Jan 2006 | A1 |
20060012567 | Sicklinger | Jan 2006 | A1 |
20060047215 | Newman et al. | Mar 2006 | A1 |
20060061468 | Ruha | Mar 2006 | A1 |
20060063993 | Yu et al. | Mar 2006 | A1 |
20060064037 | Shalon | Mar 2006 | A1 |
20060073807 | Baker et al. | Apr 2006 | A1 |
20060075257 | Martis et al. | Apr 2006 | A1 |
20060084878 | Banet et al. | Apr 2006 | A1 |
20060084879 | Nazarian et al. | Apr 2006 | A1 |
20060122520 | Banet et al. | Jun 2006 | A1 |
20060123885 | Yates et al. | Jun 2006 | A1 |
20060140425 | Berg et al. | Jun 2006 | A1 |
20060142665 | Garay et al. | Jun 2006 | A1 |
20060202816 | Crump et al. | Sep 2006 | A1 |
20060205083 | Zhao | Sep 2006 | A1 |
20060206014 | Ariav | Sep 2006 | A1 |
20060210058 | Kock et al. | Sep 2006 | A1 |
20060211922 | Al-Ali et al. | Sep 2006 | A1 |
20060211924 | Dalke et al. | Sep 2006 | A1 |
20060212316 | Jackson | Sep 2006 | A1 |
20060217598 | Miyajima et al. | Sep 2006 | A1 |
20060217988 | Sukkar | Sep 2006 | A1 |
20060224059 | Swedlow et al. | Oct 2006 | A1 |
20060226991 | Rivas | Oct 2006 | A1 |
20060240558 | Zhao | Oct 2006 | A1 |
20060246342 | MacPhee | Nov 2006 | A1 |
20060251277 | Cho | Nov 2006 | A1 |
20060251334 | Oba et al. | Nov 2006 | A1 |
20060252999 | Devaul et al. | Nov 2006 | A1 |
20060264730 | Stivoric et al. | Nov 2006 | A1 |
20060292533 | Selod | Dec 2006 | A1 |
20060293839 | Stankieiwcz et al. | Dec 2006 | A1 |
20060293921 | McCarthy et al. | Dec 2006 | A1 |
20070004449 | Sham | Jan 2007 | A1 |
20070004969 | Kong et al. | Jan 2007 | A1 |
20070015992 | Filkins et al. | Jan 2007 | A1 |
20070021206 | Sunnen | Jan 2007 | A1 |
20070027367 | Oliver et al. | Feb 2007 | A1 |
20070027399 | Chou | Feb 2007 | A1 |
20070036383 | Romero | Feb 2007 | A1 |
20070043304 | Katayama | Feb 2007 | A1 |
20070050215 | Kil et al. | Mar 2007 | A1 |
20070060800 | Drinan et al. | Mar 2007 | A1 |
20070060819 | Altschuler et al. | Mar 2007 | A1 |
20070063850 | Devaul et al. | Mar 2007 | A1 |
20070082789 | Nissila et al. | Apr 2007 | A1 |
20070083092 | Rippo et al. | Apr 2007 | A1 |
20070083095 | Rippo et al. | Apr 2007 | A1 |
20070088221 | Stahmann | Apr 2007 | A1 |
20070093702 | Yu et al. | Apr 2007 | A1 |
20070106167 | Kinast | May 2007 | A1 |
20070112273 | Rogers | May 2007 | A1 |
20070112277 | Fischer et al. | May 2007 | A1 |
20070112598 | Heckerman et al. | May 2007 | A1 |
20070116314 | Grilliot et al. | May 2007 | A1 |
20070118054 | Oliver et al. | May 2007 | A1 |
20070123763 | Al-Ali et al. | May 2007 | A1 |
20070135717 | Uenishi et al. | Jun 2007 | A1 |
20070135866 | Baker et al. | Jun 2007 | A1 |
20070159926 | Prstojevich | Jul 2007 | A1 |
20070165872 | Bridger et al. | Jul 2007 | A1 |
20070167850 | Russell et al. | Jul 2007 | A1 |
20070179739 | Donofrio et al. | Aug 2007 | A1 |
20070191718 | Nakamura | Aug 2007 | A1 |
20070197878 | Shklarski | Aug 2007 | A1 |
20070197881 | Wolf et al. | Aug 2007 | A1 |
20070213020 | Novac | Sep 2007 | A1 |
20070230714 | Armstrong | Oct 2007 | A1 |
20070233403 | Alwan et al. | Oct 2007 | A1 |
20070265097 | Havukainen | Nov 2007 | A1 |
20070265508 | Sheikhzadeh-Nadjar | Nov 2007 | A1 |
20070270667 | Coppi et al. | Nov 2007 | A1 |
20070270671 | Gal | Nov 2007 | A1 |
20070273504 | Tran | Nov 2007 | A1 |
20070276270 | Tran | Nov 2007 | A1 |
20070293781 | Sims et al. | Dec 2007 | A1 |
20070299330 | Couronne et al. | Dec 2007 | A1 |
20080001735 | Tran | Jan 2008 | A1 |
20080004536 | Baxi et al. | Jan 2008 | A1 |
20080004904 | Tran | Jan 2008 | A1 |
20080015424 | Bernreuter | Jan 2008 | A1 |
20080039731 | McCombie et al. | Feb 2008 | A1 |
20080076972 | Dorogusker et al. | Mar 2008 | A1 |
20080081963 | Naghavi et al. | Apr 2008 | A1 |
20080081972 | Debreczeny | Apr 2008 | A1 |
20080086533 | Neuhauser et al. | Apr 2008 | A1 |
20080096726 | Riley et al. | Apr 2008 | A1 |
20080106404 | Joslin et al. | May 2008 | A1 |
20080114220 | Banet et al. | May 2008 | A1 |
20080132798 | Hong et al. | Jun 2008 | A1 |
20080133699 | Craw et al. | Jun 2008 | A1 |
20080141301 | Azzaro et al. | Jun 2008 | A1 |
20080146890 | LeBoeuf et al. | Jun 2008 | A1 |
20080146892 | LeBoeuf et al. | Jun 2008 | A1 |
20080154098 | Morris et al. | Jun 2008 | A1 |
20080154105 | Lemay | Jun 2008 | A1 |
20080165017 | Schwartz | Jul 2008 | A1 |
20080170600 | Sattler et al. | Jul 2008 | A1 |
20080171945 | Dotter | Jul 2008 | A1 |
20080177162 | Bae et al. | Jul 2008 | A1 |
20080187163 | Goldstein et al. | Aug 2008 | A1 |
20080200774 | Luo | Aug 2008 | A1 |
20080203144 | Kim | Aug 2008 | A1 |
20080221461 | Zhou et al. | Sep 2008 | A1 |
20080249594 | Dietrich | Oct 2008 | A1 |
20080287752 | Stroetz et al. | Nov 2008 | A1 |
20080298624 | Jeong et al. | Dec 2008 | A1 |
20090005662 | Petersen et al. | Jan 2009 | A1 |
20090006457 | Stivoric et al. | Jan 2009 | A1 |
20090010461 | Klinghult et al. | Jan 2009 | A1 |
20090010556 | Uchibayashi et al. | Jan 2009 | A1 |
20090024004 | Yang | Jan 2009 | A1 |
20090030350 | Yang et al. | Jan 2009 | A1 |
20090034748 | Sibbald | Feb 2009 | A1 |
20090048498 | Riskey | Feb 2009 | A1 |
20090054751 | Babashan et al. | Feb 2009 | A1 |
20090054752 | Jonnalagadda et al. | Feb 2009 | A1 |
20090069645 | Nielsen et al. | Mar 2009 | A1 |
20090082994 | Schuler et al. | Mar 2009 | A1 |
20090088611 | Buschmann | Apr 2009 | A1 |
20090093687 | Telfort et al. | Apr 2009 | A1 |
20090097681 | Puria et al. | Apr 2009 | A1 |
20090105548 | Bart | Apr 2009 | A1 |
20090105556 | Fricke et al. | Apr 2009 | A1 |
20090112071 | LeBoeuf et al. | Apr 2009 | A1 |
20090131761 | Moroney, III et al. | May 2009 | A1 |
20090131764 | Lee et al. | May 2009 | A1 |
20090171221 | Liao et al. | Jul 2009 | A1 |
20090175456 | Johnson | Jul 2009 | A1 |
20090177097 | Ma et al. | Jul 2009 | A1 |
20090214060 | Chuang et al. | Aug 2009 | A1 |
20090221888 | Wijesiriwardana | Sep 2009 | A1 |
20090227853 | Wijesiriwardana | Sep 2009 | A1 |
20090240125 | Such et al. | Sep 2009 | A1 |
20090253992 | Van Der Loo | Oct 2009 | A1 |
20090253996 | Lee et al. | Oct 2009 | A1 |
20090264711 | Schuler et al. | Oct 2009 | A1 |
20090268911 | Singh | Oct 2009 | A1 |
20090270698 | Shioi et al. | Oct 2009 | A1 |
20090281435 | Ahmed et al. | Nov 2009 | A1 |
20090287067 | Dorogusker et al. | Nov 2009 | A1 |
20090299215 | Zhang | Dec 2009 | A1 |
20100004517 | Bryenton et al. | Jan 2010 | A1 |
20100004860 | Chernoguz et al. | Jan 2010 | A1 |
20100022861 | Cinbis et al. | Jan 2010 | A1 |
20100045663 | Chen et al. | Feb 2010 | A1 |
20100100013 | Hu et al. | Apr 2010 | A1 |
20100113948 | Yang et al. | May 2010 | A1 |
20100168531 | Shaltis et al. | Jul 2010 | A1 |
20100172510 | Juvonen | Jul 2010 | A1 |
20100172522 | Mooring et al. | Jul 2010 | A1 |
20100179389 | Moroney et al. | Jul 2010 | A1 |
20100185105 | Baldinger | Jul 2010 | A1 |
20100217102 | LeBoeuf et al. | Aug 2010 | A1 |
20100217103 | Abdul-Hafiz et al. | Aug 2010 | A1 |
20100222655 | Starr et al. | Sep 2010 | A1 |
20100228315 | Nielsen | Sep 2010 | A1 |
20100234714 | Mercier et al. | Sep 2010 | A1 |
20100268056 | Picard et al. | Oct 2010 | A1 |
20100274100 | Behar et al. | Oct 2010 | A1 |
20100274109 | Hu et al. | Oct 2010 | A1 |
20100292589 | Goodman | Nov 2010 | A1 |
20100298653 | McCombie et al. | Nov 2010 | A1 |
20110028810 | Van Slyke et al. | Feb 2011 | A1 |
20110028813 | Watson et al. | Feb 2011 | A1 |
20110081037 | Oh et al. | Apr 2011 | A1 |
20110105869 | Wilson et al. | May 2011 | A1 |
20110112382 | Li et al. | May 2011 | A1 |
20110130638 | Raridan, Jr. | Jun 2011 | A1 |
20110142371 | King et al. | Jun 2011 | A1 |
20110288379 | Wu | Nov 2011 | A1 |
20120030547 | Raptis et al. | Feb 2012 | A1 |
20120039493 | Rucker et al. | Feb 2012 | A1 |
20120095303 | He | Apr 2012 | A1 |
20120156933 | Kreger et al. | Jun 2012 | A1 |
20120179011 | Moon et al. | Jul 2012 | A1 |
20120197093 | LeBoeuf et al. | Aug 2012 | A1 |
20120277548 | Burton | Nov 2012 | A1 |
20130053661 | Alberth et al. | Feb 2013 | A1 |
20130063550 | Ritchey et al. | Mar 2013 | A1 |
20130072765 | Kahn et al. | Mar 2013 | A1 |
20130131519 | LeBoeuf et al. | May 2013 | A1 |
20130218588 | Kehr et al. | Aug 2013 | A1 |
20130245387 | Patel | Sep 2013 | A1 |
20130336495 | Burgett et al. | Dec 2013 | A1 |
20140051940 | Messerschmidt | Feb 2014 | A1 |
20140052567 | Bhardwaj et al. | Feb 2014 | A1 |
20140073486 | Ahmed et al. | Mar 2014 | A1 |
20140100432 | Golda et al. | Apr 2014 | A1 |
20140127996 | Park et al. | May 2014 | A1 |
20140203797 | Stivoric et al. | Jul 2014 | A1 |
20140219467 | Kurtz | Aug 2014 | A1 |
20140221777 | Betts | Aug 2014 | A1 |
20140236531 | Carter | Aug 2014 | A1 |
20140275852 | Hong et al. | Sep 2014 | A1 |
20140323880 | Ahmed et al. | Oct 2014 | A1 |
20140378844 | Fei | Dec 2014 | A1 |
20160287108 | Wei et al. | Oct 2016 | A1 |
20170034615 | Mankodi et al. | Feb 2017 | A1 |
Number | Date | Country |
---|---|---|
101212927 | Jul 2008 | CN |
201438747 | Apr 2010 | CN |
3910749 | Oct 1990 | DE |
1 297 784 | Apr 2003 | EP |
1 480 278 | Nov 2004 | EP |
2 077 091 | Jul 2009 | EP |
2 182 839 | Oct 2011 | EP |
2 408 209 | May 2005 | GB |
2 411 719 | Sep 2005 | GB |
7-241279 | Sep 1995 | JP |
9-253062 | Sep 1997 | JP |
9-299342 | Nov 1997 | JP |
2000-116611 | Apr 2000 | JP |
2001-025462 | Jan 2001 | JP |
20030159221 | Jun 2003 | JP |
2004-513750 | May 2004 | JP |
2004-283523 | Oct 2004 | JP |
2005-040261 | Feb 2005 | JP |
2005-270544 | Oct 2005 | JP |
2007-044203 | Feb 2007 | JP |
2007-185348 | Jul 2007 | JP |
2008-136556 | Jun 2008 | JP |
2008-279061 | Nov 2008 | JP |
2009-153664 | Jul 2009 | JP |
2010-526646 | Aug 2010 | JP |
2014-068733 | Apr 2014 | JP |
20-0204510 | Nov 2000 | KR |
WO 0024064 | Apr 2000 | WO |
WO 2000047108 | Aug 2000 | WO |
WO 0108552 | Feb 2001 | WO |
WO 0217782 | Mar 2002 | WO |
WO 2005010568 | Feb 2005 | WO |
WO 2005015163 | Feb 2005 | WO |
WO 2005020121 | Mar 2005 | WO |
WO 2005036212 | Apr 2005 | WO |
WO 2005074550 | Aug 2005 | WO |
WO 2005110238 | Nov 2005 | WO |
WO 2006009830 | Jan 2006 | WO |
WO 2006067690 | Jun 2006 | WO |
WO 2007012931 | Feb 2007 | WO |
WO 2007053146 | May 2007 | WO |
WO 2008141306 | Nov 2008 | WO |
WO 2011127063 | Oct 2011 | WO |
WO 2013038296 | Mar 2013 | WO |
Entry |
---|
Anpo et al. “Photocatalytic Reduction of Co2 With H2O on Titanium Oxides Anchored within Micropores of Zeolites: Effects of the Structure of the Active Sites and the Addition of Pt” J. Phys. Chem. B, 101:2632-2636 (1997). |
Bârsan et al. “Understanding the fundamental principles of metal oxide based gas sensors; the example of CO sensing with SnO2 sensors in the presence of humidity” Journal of Physics: Condensed Matter 15:R813-R839 (2003). |
Bott “Electrochemistry of Semiconductors” Current Separations 17(3):87-91 (1998). |
European Search Report corresponding to European Application No. 07862660.3 dated Apr. 25, 2012; 7 pages. |
Fitrainer “The Only Trainer You Need”; http://itami.com; Downloaded Feb. 26, 2010; © 2008 FiTriainer™; 2 pages. |
International Search Report and Written Opinion of the International Searching Authority, corresponding to PCT/US2007/025114, dated May 13, 2008. |
Martins et al. “Zinc oxide as an ozone sensor” Journal of Applied Physics 96(3):1398-1408 (2004). |
Saladin et al. “Photosynthesis of CH4 at a TiO2 Surface from Gaseous H2O and CO2” J. Chem. Soc., Chem. Commun. 533-534 (1995). |
Skubal et al. “Detection and identification of gaseous organics using a TiO2 sensor” Journal of Photochemistry and Photobiology A: Chemistry 148:103-108 (2002). |
Skubal et al. “Monitoring the Electrical Response of Photoinduced Organic Oxideation on TiO2 Surfaces” Manuscript submitted Oct. 2000 to SPIE Intl. Symposium on Environment & Industrial Sensing, Boston, MA, Nov. 5-8, 2000, sponsored by SPIE, 10 pp. |
Zhang et al. “Development of Chemical Oxygen Demand On-Line Monitoring System Based on a Photoelectrochemical Degradation Principle” Environ. Sci. Technol., 40(7):2363-2368 (2006). |
“U.S. Army Fitness Training Handbook” by the Department of the Army, 2003, The Lyons Press. p. 17. |
“Warfighter Physiological and Environmental Monitoring: A Study for the U.S. Army Research Institute in Environmental Medicine and the Soldier Systems Center”, Massachusetts Institute of Technology Lincoln Laboratory, Final Report, Nov. 1, 2004, prepared for the U.S. Army under Air Force Contract F19628-00-C-0002; approved for public release. |
Colligan, M. J. et al. in “The psychological effects of indoor air pollution”, Bulletin of the New York Academy of Medicine, vol. 57, No. 10, Dec. 1981, p. 1014-1026. |
De Paula Santos, U. et al, in “Effects of air pollution on blood pressure and heart rate variability: a panel study of vehicular traffic controllers in the city of Sao Paulo, Brazil”, European Heart Journal (2005) 26, 193-200. |
Ebert, T. et al., “Influence of Hydration Status on Thermoregulation and Cycling Hill Climbing,” Med. Sci. Sport Exerc. vol. 39, No. 2, pp. 323-329, 2007. |
Fleming et al., “A Comparison of Signal Processing Techniques for the Extraction of Breathing Rate from the Photopethysmorgram,” World Academy of Science, Engineering and Technology, vol. 30. Oct. 2007, pp. 276-280. |
Geladas et al., “Effect of cold air inhalation on core temperature in exercising subjects under stress,” The American Physiological Society, pp. 2381-2387, 1988. |
Gold, D.R. et al. in “Ambient Pollution and Heart Rate Variability”, Circulation 2000, 101:1267-1273. |
International Search Report corresponding to International Patent Application No. PCT/US2017/046446, dated Jan. 14, 2013, 3 pages. |
International Search Report and Written Opinion of the International Searching Authority, corresponding to PCT/US2012/0948079, dated Oct. 9, 2012. |
International Search Report Corresponding to International Application No. PCT/US2012/022634, dated Aug. 22, 2012, 9 pages. |
Maomao et al., “Mobile Context-Aware Game for the Next Generation,” 2nd International Conference on Application and Development of Computer Games ADCOG 2003, p. 78-81. |
Maughan, R.J., “Impact of mild dehydration on wellness and on exercise performance,” European Journal of Clinical Nutrition, 57, Suppl. 2, pp. S19-S23, 2003. |
Maughan et al., “Exercise, Heat, Hydration and the Brain,” Journal of the American College of Nutrition, vol. 26, No. 5, pp. 604S-612S, 2007. |
Mostardi, R., et al., “The effect of increased body temperature due to exercise on the heart rate and the maximal aerobic power,” Europ. J. Appl. Physiol, 33, pp. 237-245, 1974. |
Nakajima et al., “Monitoring of heart and respiratory rates by photoplethyusmography using a digital filtering technique,” Med. Eng. Phys., vol. 18, No. 5, Jul. 1996, pp. 365-372. |
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated Jul. 30, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/021936. |
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated Aug. 26, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/021629. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Search Authority dated Sep. 16, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/024922. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Search Authority dated Sep. 27, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/025216. |
Notification of Transmittal of The International Search Report and The Written Opinion of the International Searching Authority, or the Declaration corresponding to International Application No. PCT/US2013/070271; dated Feb. 26, 2014; 13 pages. |
Shorten et al., “Acute effect of environmental temperature during exercise on subsequent energy intake in active men,” Am. J Clin. Nutr. 90, pp. 1215-1221, 2009. |
Thompson, M.W., “Cardiovascular drift and critical core temperature: factors limiting endurance performance in the heat?” J. Exerc. Sci. Fit, vol. 4, No. 1, pp. 15-24, 2006. |
Edmison et al., “E-Textile Based Automatic Activity Diary for Medical Annotation and Analysis,” Proc. BSN 2006 Int. Workshop Wearable Implantable Body Sensor Netw. (2006), pp. 131-145, Apr. 3-5, 2006. |
Gibbs et al., “Reducing Motion Artifact in Wearable Bio-Sensors Using MEMS Accelerometers for Active Noise Cancellation,” 2005 American Control Conference, Jun. 8-10, 2005, Portland, OR, USA, pp. 1581-1586. |
International Search Report corresponding to International Patent Application No. PCT/US2014/012909, dated May 13, 2014, 3 pages. |
Notification Concerning Transmittal of International Preliminary Report on Patentability, PCT/US2014/012909, dated Jul. 28, 2015. |
Wood et al., “Active Motion Artifact Reduction for Wearable Sensors Using Laguerre Expansion and Signal Separation,” Proceedings of the 2005 IEEE Engineering in Medicine and Biology, 27th Annual Conference, Shanghai, China, Sep. 1-4, 2005, pp. 3571-3574. |
International Preliminary Report on Patentability, PCT/US2014/012940, dated Jun. 17, 2015, 23 pages. |
International Search Report and Written Opinion of the International Searching Authority, corresponding to International Patent Application No. PCT/US2014/012940, dated Oct. 16, 2014, 13 pages. |
Communication pursuant to Article 94(3) EPC, European Patent Application No. 13863449.8, dated Nov. 5, 2015, 7 pages. |
Communication pursuant to Article 94(3) EPC, European Patent Application No. 14743615.8, dated Dec. 23, 2015, 7 pages. |
Communication pursuant to Article 94(3) EPC, European Patent Application No. 14743839.4, dated Dec. 23, 2015, 6 pages. |
European Search Report, EP Application No. 13863449.8, dated Oct. 19, 2015, 3 pages. |
European Search Report, EP Application No. 14743615.8, dated Oct. 12, 2015, 3 pages. |
European Search Report, EP Application No. 14743839.4, dated Oct. 12, 2015, 3 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/014562, dated Oct. 28, 2015. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/042636, dated Oct. 29, 2015. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/042015, dated Oct. 29, 2015. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/042035, dated Oct. 29, 2015. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/046079, dated Dec. 29, 2015. |
Communication pursuant to Article 94(3) EPC, European Patent Application No. 12820308.0, dated Feb. 3, 2016, 5 pages. |
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated May 26, 2016 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2016/019126. |
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated May 26, 2016 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2016/019132. |
Asada, et al., “Mobile Monitoring with Wearable Photoplethysmographic Biosensors,” IEEE Engineering in Medicine and Biology Magazine, May/Jun. 2003, pp. 28-40. |
Bifulco et al., “Bluetooth Portable Device for Continuous ECG and Patient Motion Monitoring During Daily Life,” Medicon 2007, IFMBE Proceedings 16, 2007, pp. 369-372. |
Brodersen et al., “In-Ear Acquisition of Vital Signs Discloses New Chances for Preventive Continuous Cardiovascular Monitoring,” 4th International Workshop on Wearable and Implantable Body Sensor Networks (BSN 2007), vol. 13 of the series IFMBE Proceedings, pp. 189-194. |
Celka et al, “Motion Resistant Earphone Located Infrared based Heart Rate Measurement Device,” Proceedings of the Second IASTED International Conference on Biomedical Engineering, Feb. 16-18, 2004, Innsbruck, Austria, pp. 582-585. |
Communication Pursuant to Article 94(3) EPC, EP 12 739 502.8, dated Jul. 19, 2016, 7 pages. |
Communication Pursuant to Article 94(3) EPC, EP 14 743 615.8, dated Jul. 19, 2016, 7 pages. |
Communication Pursuant to Article 94(3) EPC, EP 14 743 839.4, dated Jul. 20, 2016, 5 pages. |
Comtois et al., “A Wearable Wireless Reflectance Pulse Oximeter for Remote Triage Applications,” 2006 IEEE, pp. 53-54. |
Comtois, Gary, W., “Implementation of Accelerometer-Based Adaptive Noise Cancellation in a Wireless Wearable Pulse Oximeter Platform for Remote Physiological Monitoring and Triage,” Thesis, Worcester Polytechnic Institute, Aug. 31, 2007, 149 pages. |
Duun et al., “A Novel Ring Shaped Photodiode for Reflectance Pulse Oximetry in Wireless Applications,” IEEE Sensors 2007 Conference, pp. 596-599. |
Geun et al., “Measurement Site and Applied Pressure Consideration in Wrist Photoplethysmography,” The 23rd International Technical Conference on Circuits/Systems, Computers and Communications, 2008, pp. 1129-1132. |
Gibbs et al., “Active motion artifact cancellation for wearable health monitoring sensors using collocated MEMS accelerometers,” Smart Structures and Materials, 2005: Sensors and Smart Structures Technologies for Civil, Mechanical, and Aerospace Systems, Proc. of SPIE, vol. 5765, pp. 811-819. |
Haahr et al., “A Wearable “Electronic Patch” for Wireless Continuous Monitoring of Chronically Diseased Patients,” Proceedings of the 5th International Workshop on Wearable and Implantable Body Sensor Networks, in conjunction with The 5th International Summer School and Symposium on Medical Devices and Biosensors, The Chinese University of Hong Kong, HKSAR, China, Jun. 1-3, 2008, pp. 66-70. |
Jiang, Honghui, “Motion-Artifact Resistant Design of Photoplethysmograph Ring Sensor for Driver Monitoring,” Thesis, Massachusetts Institute of Technology, Feb. 2004, 62 pages. |
Kuzmina et al., “Compact multi-functional skin spectrometry set-up,” Advanced Optical Materials, Technologies, and Devices, Proc. of SPIE, vol. 6596, 2007, pp. 65960T-1 to 65960T-6. |
Lee et al, “Respiratory Rate Detection Algorithms by Photoplethysmography Signal Processing,” 30th Annual International IEEE EMBS Conference, Vancouver, British Columbia, Canada, Aug. 20-24, 2008, pp. 1140-1143. |
Lindberg et al., “Monitoring of respiratory and heart rates using a fibre-optic sensor,” Med Biol Eng Comput, Sep. 1992, vol. 30, No. 5, pp. 533-537. |
Luprano, Jean, “Sensors and Parameter Extraction by Wearable Systems: Present Situation and Future,” pHealth 2008, May 21, 2008, 29 pages. |
Lygouras et al., “Optical-Fiber Finger Photo-Plethysmograph Using Digital Techniques,” IEEE Sensors Journal, vol. 2, No. 1, Feb. 2002, pp. 20-25. |
Maguire et al., “The Design and Clinical Use of a Reflective Brachial Photoplethysmograph,” Technical Report NUIM/SS/—/2002/04, Submitted Apr. 2002, Signals and Systems Research Group, National University of Ireland, Maynooth, Co. Kildare, Ireland, 13 pages. |
Mendelson et al., “Measurement Site and Photodetector Size Considerations in Optimizing Power Consumption of a Wearable Reflectance Pulse Oximeter,” Proceedings of the 25th Annual International Conference of the IEEE EMBS, Cancun, Mexico, Sep. 17-21, 2003, pp. 3016-3019. |
Mendelson et al., “Noninvasive Pulse Oximetry Utilizing Skin Reflectance Photoplethysmography,” IEEE Transactions on Biomedical Engineering, vol. 35, No. 10, Oct. 1988, pp. 798-805. |
Poh et al., “Motion Tolerant Magnetic Earring Sensor and Wireless Earpiece for Wearable Photoplethysmography,” IEEE Transactions on Information Technology in Biomedicine, vol. 14, No. 3, May 2010, pp. 786-794. |
Renevey et al., “Wrist-Located Pulse Detection Using IR Signals, Activity and Nonlinear Artifact Cancellation,” IEEE EMBS, 2001, 4 pages. |
Rhee et al., “Artifact-Resistant Power-Efficient Design of Finger-Ring Plethysmographic Sensors,” IEEE Transactions on Biomedical Engineering, vol. 48, No. 7, Jul. 2001, pp. 795-805. |
Shaltis, Phillip Andrew, Analysis and Validation of an Artifact Resistant Design for Oxygen Saturation Measurement Using Photo Plethysmographic Ring Sensors, Thesis, Massachusetts Institute of Technology, Jun. 2004, 103 pages. |
Shin et al., “A Novel Headset with a Transmissive PPG Sensor for Heart Rate Measurement,” ICBME 2008, Proceedings 23, 2009, pp. 519-522. |
Spigulis et al, “Wearable wireless photoplethysmography sensors,” Proc. of SPIE, vol. 6991, 2008, pp. 69912O-1 to 69912O-7. |
Takatani et al., “Optical Oximetry Sensors for Whole Blood and Tissue, IEEE Engineering in Medicine and Biology,” Jun./Jul. 1994, pp. 347-357. |
Vogel et al., “A System for Assessing Motion Artifacts in the Signal of a Micro-Optic In-Ear Vital Signs Sensor,” 30th Annual International IEEE EMBS Conference, Vancouver, British Columbia, Canada, Aug. 20-24, 2008. |
Vogel et al., “In-Ear Heart Rate Monitoring Using a Micro-Optic Reflective Sensor,” Proceedings of the 29th Annual International Conference of the IEEE EMBS Cite Internationale, Lyon, France, Aug. 23-26, 2007, pp. 1375-1378. |
Wang et al., “Multichannel Reflective PPG Earpiece Sensor With Passive Motion Cancellation,” IEEE Transactions on Biomedical Circuits and Systems, vol. 1, No. 4, Dec. 2007, pp. 235-241. |
Wang et al., “Reflective Photoplethysmograph Earpiece Sensor for Ubiquitous Heart Rate Monitoring,” 4th International Workshop on Wearable and Implantable Body Sensor Networks, 2007, vol. 13 of the series IFMBE Proceedings, pp. 179-183. |
Wei et al. “A New Wristband Wearable Sensor Using Adaptive Reduction Filter to Reduce Motion Artifact,” Proceedings of the 5th International Conference on Information Technology and Application in Biomedicine, in conjunction with The 2nd International Symposium & Summer School on Biomedical and Health Engineering, Shenzhen, China, May 30-31, 2008, pp. 278-281. |
Wood, Levi Benjamin, “Motion Artifact Reduction for Wearable Photoplethysmogram Sensors Using Micro Accelerometers and Laguerre Series Adaptive Filters,” Thesis, Massachusetts Institute of Technology, Jun. 2008, 74 pages. |
Han et al., “Artifacts in wearable photoplethysmographs during daily life motions and their reduction with least mean square based active noise cancellation method,” Computers in Biology and Medicine, 42, 2012, pp. 387-393. |
Extended European Search Report, EP Application No. 16164775.5 dated Sep. 13, 2016, 7 pages. |
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2016/041842, dated Oct. 21, 2016, 5 pages. |
Notification of Transmittal of International Preliminary Report on Patentability, PCT/US2015/041562, dated Oct. 20, 2016, 14 pages. |
Notification of Transmittal of International Preliminary Report on Patentability, PCT/US2015/042636, dated Oct. 20, 2016, 7 pages. |
Notification of Transmittal of International Preliminary Report on Patentability, PCT/US2015/042015, dated Oct. 20, 2016, 10 pages. |
Notification of Transmittal of International Preliminary Report on Patentability, PCT/US2015/042035, dated Oct. 20, 2016, 8 pages. |
Notification of Transmittal of International Preliminary Report on Patentability, PCT/US2015/046079, dated Oct. 20, 2016, 10 pages. |
Communication with Supplementary European Search Report, European Application No. 15830336.2, dated Jun. 7, 2017, 8 pp. |
Comtois et al., “A Comparative Evaluation of Adaptive Noise Cancellation Algorithms for Minimizing Motion Artifacts in a Forehead-Mounted Wearable Pulse Oximeter”, Proceedings of the 29th Annual International Conference of the IEEE EMBS, Lyon, France, Aug. 23-26, 2007, pp. 1528-1531. |
Han et al. “Development of a wearable health monitoring device with motion artifact reduced algorithm” International Conference on Control, Automation and Systems 2007 (ICCAS 2007), Seoul, Korea, Oct. 17-20, 2007, pp. 1581-1584. |
Lee et al., “A Mobile Care System With Alert Mechanism”, IEEE Transactions On Information Technology In Biomedicine, vol. 11, No. 5, Sep. 2007, pp. 507-517. |
Webster, J. G. Design of Pulse Oximeters. IOP Publishing Ltd., 1997, Cover page, pp. i-xvi, pp. 34-159. |
Number | Date | Country | |
---|---|---|---|
20180184916 A1 | Jul 2018 | US |
Number | Date | Country | |
---|---|---|---|
60905761 | Mar 2007 | US | |
60876128 | Dec 2006 | US | |
60875606 | Dec 2006 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14063669 | Oct 2013 | US |
Child | 15910711 | US | |
Parent | 11811844 | Jun 2007 | US |
Child | 14063669 | US |