Physiological monitoring apparatus and networks

Information

  • Patent Grant
  • 11412938
  • Patent Number
    11,412,938
  • Date Filed
    Tuesday, April 10, 2018
    6 years ago
  • Date Issued
    Tuesday, August 16, 2022
    2 years ago
Abstract
Wearable apparatus for monitoring various physiological and environmental factors are provided. Real-time, noninvasive health and environmental monitors include a plurality of compact sensors integrated within small, low-profile devices, such as earpiece modules. Physiological and environmental data is collected and wirelessly transmitted into a wireless network, where the data is stored and/or processed.
Description
FIELD OF THE INVENTION

The present invention relates generally to health and environmental monitors and, more particularly, to wireless health and environment monitors.


BACKGROUND OF THE INVENTION

There is growing market demand for personal health and environmental monitors, for example, for gauging overall health and metabolism during exercise, athletic training, dieting, and physical therapy. However, traditional health monitors and environmental monitors may be bulky, rigid, and uncomfortable—generally not suitable for use during daily physical activity. There is also growing interest in generating and comparing health and environmental exposure statistics of the general public and particular demographic groups. For example, collective statistics enable the healthcare industry and medical community to direct healthcare resources to where they are most highly valued. However, methods of collecting these statistics may be expensive and laborious, often utilizing human-based recording/analysis steps at multiple sites.


As such, improved ways of collecting, storing and analyzing personal health and environmental information are needed. In addition, improved ways of distributing raw and analyzed personal health and environmental information are desirable to support efforts to enhance healthcare quality and reduce costs.


SUMMARY

In view of the above discussion, apparatus for monitoring various physiological and environmental factors are provided. According to some embodiments of the present invention, real-time, noninvasive health and environmental monitors include a plurality of compact sensors integrated within small, low-profile devices. Physiological and environmental data is collected and wirelessly transmitted into a wireless network, where the data is stored and/or processed.


In some embodiments of the invention, an earpiece functions as a physiological monitor, an environmental monitor, and a wireless personal communicator. The earpiece can take advantage of commercially available open-architecture wireless paradigms, such as Bluetooth®, Wi-Fi, or ZigBee. In some embodiments, a small, compact earpiece contains at least one microphone and one speaker, and is configured to transmit information wirelessly to a recording device such as, for example, a cell phone, a personal digital assistant (PDA), and/or a computer. The earpiece contains a plurality of sensors for monitoring personal health and environmental exposure. Health and environmental information, sensed by the sensors is transmitted wirelessly, in real-time, to a recording device, capable of processing and organizing the data into meaningful displays, such as charts. In some embodiments, an earpiece user can monitor health and environmental exposure data in real-time, and may also access records of collected data throughout the day, week, month, etc., by observing charts and data through an audio-visual display.


In some embodiments, an earpiece can integrate personal physiological and environmental exposure information with biofeedback and personal entertainment. In other embodiments of the present invention, earpiece monitor devices enable a variety of networks, applications, games, and business methods.


In some embodiments of the present invention, a monitoring apparatus includes a housing configured to be attached to the body of a person, one or more physiological sensors and one or more environmental sensors supported by (within and/or on) the housing. Each physiological sensor is configured to detect and/or measure physiological information from the person, and each environmental sensor is configured to detect and/or measure environmental conditions in a vicinity of the person wearing the apparatus. The apparatus also includes a signal processor that is configured to receive and process signals produced by the physiological and environmental sensors. A wireless transmitter is responsive to the signal processor and is configured to wirelessly transmit physiological and environmental sensor signals as processed by the signal processor from the signal processor to a remote terminal in real-time.


Each physiological sensor is configured to detect and/or measure one or more of the following types of physiological information: heart rate, pulse rate, breathing rate, blood flow, heartbeat signatures, cardio-pulmonary health, organ health, metabolism, electrolyte type and/or concentration, physical activity, caloric intake, caloric metabolism, blood metabolite levels or ratios, blood pH level, physical and/or psychological stress levels and/or stress level indicators, drug dosage and/or dosimetry, physiological drug reactions, drug chemistry, biochemistry, position and/or balance, body strain, neurological functioning, brain activity, brain waves, blood pressure, cranial pressure, hydration level, auscultatory information, auscultatory signals associated with pregnancy, physiological response to infection, skin and/or core body temperature, eye muscle movement, blood volume, inhaled and/or exhaled breath volume, physical exertion, exhaled breath physical and/or chemical composition, the presence and/or identity and/or concentration of viruses and/or bacteria, foreign matter in the body, internal toxins, heavy metals in the body, anxiety, fertility, ovulation, sex hormones, psychological mood, sleep patterns, hunger and/or thirst, hormone type and/or concentration, cholesterol, lipids, blood panel, bone density, organ and/or body weight, reflex response, sexual arousal, mental and/or physical alertness, sleepiness, auscultatory information, response to external stimuli, swallowing volume, swallowing rate, sickness, voice characteristics, voice tone, voice pitch, voice volume, vital signs, head tilt, allergic reactions, inflammation response, auto-immune response, mutagenic response, DNA, proteins, protein levels in the blood, water content of the blood, pheromones, internal body sounds, digestive system functioning, cellular regeneration response, healing response, stem cell regeneration response


Each environmental sensor is configured to detect and/or measure one or more of the following types of environmental information: climate, humidity, temperature, pressure, barometric pressure, soot density, airborne particle density, airborne particle size, airborne particle shape, airborne particle identity, volatile organic chemicals (VOCs), hydrocarbons, polycyclic aromatic hydrocarbons (PAHs), carcinogens, toxins, electromagnetic energy, optical radiation, X-rays, gamma rays, microwave radiation, terahertz radiation, ultraviolet radiation, infrared radiation, radio waves, atomic energy alpha particles, atomic energy beta-particles, gravity, light intensity, light frequency, light flicker, light phase, ozone, carbon monoxide, carbon dioxide, nitrous oxide, sulfides, airborne pollution, foreign material in the air, viruses, bacteria, signatures from chemical weapons, wind, air turbulence, sound and/or acoustical energy, ultrasonic energy, noise pollution, human voices, animal sounds, diseases expelled from others, exhaled breath and/or breath constituents of others, toxins from others, pheromones from others, industrial and/or transportation sounds, allergens, animal hair, pollen, exhaust from engines, vapors and/or fumes, fuel, signatures for mineral deposits and/or oil deposits, snow, rain, thermal energy, hot surfaces, hot gases, solar energy, hail, ice, vibrations, traffic, the number of people in a vicinity of the person, coughing and/or sneezing sounds from people in the vicinity of the person, loudness and/or pitch from those speaking in the vicinity of the person.


In some embodiments, the signal processor is configured to process signals produced by the physiological and environmental sensors into signals that can be heard and/or viewed by the person wearing the apparatus. In some embodiments, the signal processor is configured to selectively extract environmental effects from signals produced by a physiological sensor and/or selectively extract physiological effects from signals produced by an environmental sensor.


In some embodiments of the present invention, a monitoring apparatus configured to be worn by a person includes a physiological sensor that is oriented in a direction towards the person and an environmental sensor that is oriented in a direction away from the person. A buffer material is positioned between the physiological sensor and environmental sensors and is configured to selectively reflect and/or absorb energy emanating from the environment and/or the person.


In some embodiments of the present invention, a monitoring apparatus may include a receiver that is configured to receive audio and/or video information from a remote terminal, and a communication module that is configured to store and/or process and/or play audio and/or video information received from the remote terminal. In some embodiments, the communication module may be configured to alert (e.g., via audible and/or visible and/or physical alerts) a person wearing the apparatus when a physiological sensor detects certain physiological information from the person and/or when an environmental sensor detects certain environmental information from the vicinity of the person. In some embodiments, the communication module is configured to audibly present vital sign information to the person wearing the apparatus. In some embodiments, the communication module may be configured to store content generated by the person.


In some embodiments of the present invention, a monitoring apparatus may include a transmitter that is configured to transmit signals produced by physiological and environmental sensors associated therewith to a gaming device. The monitoring apparatus may also be configured to receive feedback regarding monitored health and environmental parameters. As such, personal health and environmental feedback can be an active component of a game.


In some embodiments, the apparatus is an earpiece module that is configured to be attached to the ear of a person, and includes a speaker, microphone, and transceiver that is electronically connected to the speaker and microphone and that permits bidirectional wireless communications between the earpiece module and a remote terminal, such as a cell phone. The transceiver (e.g., a Bluetooth®, Wi-Fi, or ZigBee transceiver) is electronically connected to the signal processor and is configured to transmit physiological and environmental sensor signals from the signal processor to the remote terminal. In some embodiments, the earpiece module may include an arm that is attached to the housing and that supports the microphone. The arm may be movable between a stored position and an extended, operative position. The arm may also include one or more physiological sensor and/or environmental sensors.


In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes a first acoustical sensor oriented in a direction towards a tympanic membrane of the ear and is configured to detect acoustical energy emanating from the tympanic membrane. A second acoustical sensor is oriented in a direction away from the person. The signal processor is configured to utilize signals produced by the second acoustical signal to extract environmental acoustical energy not emanating from the tympanic membrane from signals produced by the first acoustical sensor. In some embodiments, the earpiece module may include an optical emitter that directs optical energy towards the tympanic membrane, and an optical detector that is configured to detect secondary optical energy emanating from the tympanic membrane. The signal processor is configured to extract selected optical energy from the secondary optical energy emanating from the tympanic membrane. The signal processor may also be configured to extract optical noise from the secondary optical energy emanating from the tympanic membrane. In some embodiments, the optical detector may include a filter configured to pass secondary optical energy at selective wavelengths.


In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes an optical detector that is configured to detect acoustically modulated blackbody IR radiation emanating from the tympanic membrane.


In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes an optical emitter that directs optical energy towards the tympanic membrane, and an optical detector configured to detect secondary optical energy emanating from the tympanic membrane. In some embodiments, the signal processor may be configured to extract selected optical energy and/or optical noise from the secondary optical energy emanating from the tympanic membrane. In some embodiments, the optical detector may include a filter configured to pass secondary optical energy at selective wavelengths.


In some embodiments of the present invention, an earpiece module that is configured to be attached to the ear of a person includes an ear hook that is configured to attach to an ear of a person. One or more physiological sensors and/or one or more environmental sensors may be supported by the ear hook. In some embodiments, the hook may include a pinna cover that is configured to contact a portion of the pinna of an ear. One or more physiological and/or environmental sensors may be supported by the pinna cover.


In some embodiments of the present invention, an earpiece module may include an arm that extends outwardly therefrom and that supports one or more physiological sensors and/or environmental sensors. For example, the arm may be configured to support physiological sensors configured to detect and/or measure jaw motion and/or arterial blood flow near the neck of a person wearing the earpiece module.


In some embodiments of the present invention, an earpiece module may include an earpiece fitting configured to be inserted near or within the ear canal of a person wearing the earpiece. The earpiece fitting may include one or more physiological sensors configured to detect information from within the ear canal.


In some embodiments of the present invention, an earpiece module may include a transmittance pulse oximeter and/or reflectance pulse oximeter. For example, the earpiece module may include an earlobe clip having a transmittance pulse oximeter and/or reflectance pulse oximeter supported thereby. As another example, the earpiece module may include a transmitter pulse oximeter and/or reflectance pulse oximeter supported at the front or back of the ear.


In some embodiments of the present invention, a monitoring apparatus is an earring. The earring may be configured to operate independently of other monitoring apparatus, such as an earpiece module, or may operate in conjunction with another monitoring apparatus. For example, an earring may include one or more physiological sensors configured to detect and/or measure physiological information from the person, and one or more environmental sensors configured to detect and/or measure environmental conditions in a vicinity of the person wearing the earring. The earring may also include a signal processor that receives and processes signals produced by the physiological and environmental sensors, and a transmitter that transmits physiological and environmental sensor signals from the signal processor to a remote terminal in real-time.


In some embodiments of the present invention, a monitoring apparatus configured to be attached to the ear of a person may include a housing containing one or more physiological and environmental sensors wherein the housing is configured to be positioned in adjacent contacting relationship with the temple of the person.


Monitoring apparatus, according to some embodiments of the present invention, may include various additional devices/features. For example, a monitoring apparatus may include an air sampling system that samples air in a vicinity of the person wearing the apparatus. In some embodiments, one or more physiological sensors in a monitoring apparatus may be configured to detect drowsiness of the person wearing the apparatus. An alarm may be provided that is configured to alert the person in response to one or more physiological sensors detecting drowsiness. In some embodiments, a monitoring apparatus may include a user interface that provides user control over one or more of the physiological and/or environmental sensors. A user interface may be provided on the monitoring apparatus or may be included on a remote device in wireless communication with the monitoring apparatus. In some embodiments, a monitoring apparatus may include a user interface that is configured to allow the person to store a time mark indicating a particular point in time.


Monitoring apparatus, according to some embodiments of the present invention, may be configured to send a signal to a remote terminal when one or more of the physiological and/or environmental sensors are turned off by a user and/or when one or more of the physiological and/or environmental sensors malfunction or fail. In some embodiments, a signal may be sent to a remote terminal when potentially erroneous data has been collected by one or more of the physiological and/or environmental sensors, such as when a person wearing a monitoring apparatus is surrounded by loud noises.


Monitoring apparatus, according to some embodiments of the present invention, may be configured to detect damage to a portion of the body of the person wearing the apparatus, and may be configured to alert the person when such damage is detected. For example, when a person is exposed to sound above a certain level that may be potentially damaging, the person is notified by the apparatus to move away from the noise source. As another example, the person may be alerted upon damage to the tympanic membrane due to loud external noises.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 is a block diagram of a telemetric earpiece module for physiological and environmental monitoring and personal communication, according to some embodiments of the present invention.



FIG. 2 is a block diagram of a telemetric network for health and environmental monitoring through portable telemetric sensor modules, such as the earpiece module of FIG. 1, according to some embodiments of the present invention.



FIG. 3 illustrates a graphical user interface for displaying data, according to some embodiments of the present invention.



FIG. 4 is a block diagram that illustrates a method of extracting physiological and environmental information using a plurality of sensors and a signal processor, according to some embodiments of the present invention.



FIG. 5 illustrates an auscultatory signal extraction technique according to the methodology illustrated in FIG. 4.



FIG. 6 illustrates an optical physiological signal extraction technique, according to some embodiments of the present invention, and wherein optical information scattered from the tympanic membrane is digitally compared with acoustical energy from the environment to generate an extracted signal containing cleaner physiological information than raw optical information scattered from the tympanic membrane.



FIG. 7 illustrates an optical source detector configuration, according to some embodiments of the present invention, for the physiological signal extraction method illustrated in FIG. 6.



FIG. 8 illustrates experimental auscultatory data obtained via the auscultatory signal extraction approach of FIG. 5.



FIG. 9 illustrates an earpiece module according to some embodiments of the present invention.



FIG. 10 is a side view of the earpiece module of FIG. 9 showing a placement of physiological sensors, according to some embodiments of the present invention.



FIG. 11 is a front view of the earpiece module of FIG. 9 showing a placement of environmental sensors, according to some embodiments of the present invention.



FIG. 12 is an exploded view of the earpiece module of FIG. 9 showing a location of various physiological sensors, according to some embodiments of the present invention.



FIG. 13 is a side view of a flexible substrate configured to place sensors in selected locations in the vicinity of the ear, according to some embodiments of the present invention.



FIGS. 14A-14B illustrates an earpiece module with an adjustable mouthpiece for monitoring physiological and environmental information near the mouth, according to some embodiments of the present invention, wherein FIG. 14A illustrates the mouthpiece in a stored position and wherein FIG. 14B illustrates the mouthpiece in an extended operative position.



FIG. 15 illustrates an earpiece module incorporating various physiological and environmental sensors, according to some embodiments of the present invention, and being worn by a user.



FIG. 16 illustrates an earpiece module according to other embodiments of the present invention that includes a temple module for physiological and environmental monitoring.



FIG. 17 illustrates a pulse-oximeter configured to be attached to an ear of a user and that may be incorporated into an earpiece module, according to some embodiments of the present invention. The illustrated pulse-oximeter is in transmission mode.



FIG. 18 illustrates a pulse-oximeter configured to be integrated into an earpiece module, according to some embodiments of the present invention. The illustrated pulse-oximeter is in reflection mode.



FIG. 19 illustrates a sensor module having a plurality of health and environmental sensors and mounted onto a Bluetooth headset module, according to some embodiments of the present invention.



FIG. 20 is a pie chart that graphically illustrates exemplary power usage of an earpiece module for monitoring health and environmental exposure, according to some embodiments of the present invention.





DETAILED DESCRIPTION

The present invention now is described more fully hereinafter with reference to the accompanying drawings, in which preferred embodiments of the invention are shown. This invention may, however, be embodied in many different forms and should not be construed as limited to the embodiments set forth herein; rather, these embodiments are provided so that this disclosure will be thorough and complete, and will fully convey the scope of the invention to those skilled in the art.


Like numbers refer to like elements throughout. In the figures, the thickness of certain lines, layers, components, elements or features may be exaggerated for clarity.


The terminology used herein is for the purpose of describing particular embodiments only and is not intended to be limiting of the invention. As used herein, the singular forms “a”, “an” and “the” are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms “comprises” and/or “comprising,” when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof. As used herein, the term “and/or” includes any and all combinations of one or more of the associated listed items.


Unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the specification and relevant art and should not be interpreted in an idealized or overly formal sense unless expressly so defined herein. Well-known functions or constructions may not be described in detail for brevity and/or clarity.


It will be understood that when an element is referred to as being “on”, “attached” to, “connected” to, “coupled” with, “contacting”, etc., another element, it can be directly on, attached to, connected to, coupled with or contacting the other element or intervening elements may also be present. In contrast, when an element is referred to as being, for example, “directly on”, “directly attached” to, “directly connected” to, “directly coupled” with or “directly contacting” another element, there are no intervening elements present. It will also be appreciated by those of skill in the art that references to a structure or feature that is disposed “adjacent” another feature may have portions that overlap or underlie the adjacent feature.


Spatially relative terms, such as “under”, “below”, “lower”, “over”, “upper” and the like, may be used herein for ease of description to describe one element or feature's relationship to another element(s) or feature(s) as illustrated in the figures. It will be understood that the spatially relative terms are intended to encompass different orientations of the device in use or operation in addition to the orientation depicted in the figures. For example, if the device in the figures is inverted, elements described as “under” or “beneath” other elements or features would then be oriented “over” the other elements or features. Thus, the exemplary term “under” can encompass both an orientation of “over” and “under”. The device may be otherwise oriented (rotated 90 degrees or at other orientations) and the spatially relative descriptors used herein interpreted accordingly. Similarly, the terms “upwardly”, “downwardly”, “vertical”, “horizontal” and the like are used herein for the purpose of explanation only unless specifically indicated otherwise.


The term “earpiece module” includes any type of device that may be attached to or near the ear of a user and may have various configurations, without limitation.


The term “real-time” is used to describe a process of sensing, processing, or transmitting information in a time frame which is equal to or shorter than the minimum timescale at which the information is needed. For example, the real-time monitoring of pulse rate may result in a single average pulse-rate measurement every minute, averaged over 30 seconds, because an instantaneous pulse rate is often useless to the end user. Typically, averaged physiological and environmental information is more relevant than instantaneous changes. Thus, in the context of the present invention, signals may sometimes be processed over several seconds, or even minutes, in order to generate a “real-time” response.


The term “monitoring” refers to the act of measuring, quantifying, qualifying, estimating, sensing, calculating, interpolating, extrapolating, inferring, deducing, or any combination of these actions. More generally, “monitoring” refers to a way of getting information via one or more sensing elements. For example, “blood health monitoring” includes monitoring blood gas levels, blood hydration, and metabolite/electrolyte levels.


The term “physiological” refers to matter or energy of or from the body of a creature (e.g., humans, animals, etc.). In embodiments of the present invention, the term “physiological” is intended to be used broadly, covering both physical and psychological matter and energy of or from the body of an organism. However, in some cases, the term “psychological” is called-out separately to emphasize aspects of physiology that are more closely tied to conscious or subconscious brain activity rather than the activity of other organs, tissues, or cells.


The term “psychosocial stress” refers to events of psychological or social origin which challenge the homeostatic state of biological systems.


The term “body” refers to the body of a person (or animal) that may utilize an earpiece module according to embodiments of the present invention. Monitoring apparatus, according to embodiments of the present invention may be worn by humans and animals.


In the following figures, earpiece modules will be illustrated and described for attachment to the ear of the human body. However, it is to be understood that embodiments of the present invention are not limited to those worn by humans. Moreover, monitoring apparatus according to embodiments of the present invention are not limited to earpiece modules and/or devices configured to be attached to or near the ear. Monitoring apparatus according to embodiments of the present invention may be worn on various parts of the body.


Some embodiments of the present invention may arise from a discovery that the ear is an ideal location on the human body for a wearable health and environmental monitor. The ear is a relatively immobile platform that does not obstruct a person's movement or vision. Devices located along the ear have access to the inner-ear canal and tympanic membrane (for measuring core body temperature), muscle tissue (for monitoring muscle tension), the pinna and earlobe (for monitoring blood gas levels), the region behind the ear (for measuring skin temperature and galvanic skin response), and the internal carotid artery (for measuring cardiopulmonary functioning). The ear is also at or near the point of exposure to: environmental breathable toxicants of interest (volatile organic compounds, pollution, etc.; noise pollution experienced by the ear; and lighting conditions for the eye. Furthermore, as the ear canal is naturally designed for transmitting acoustical energy, the ear provides an optimal location for monitoring internal sounds, such as heartbeat, breathing rate, and mouth motion.


Bluetooth-enabled and/or other personal communication earpiece modules may be configured to incorporate physiological and/or environmental sensors, according to some embodiments of the present invention. Bluetooth earpiece modules are typically lightweight, unobtrusive devices that have become widely accepted socially. Moreover, Bluetooth earpiece modules are cost effective, easy to use, and are often worn by users for most of their waking hours while attending or waiting for cell phone calls. Bluetooth earpiece modules configured according to embodiments of the present invention are advantageous because they provide a function for the user beyond health monitoring, such as personal communication and multimedia applications, thereby encouraging user compliance. Exemplary physiological and environmental sensors that may be incorporated into a Bluetooth or other type of earpiece module include, but are not limited to accelerometers, auscultatory sensors, pressure sensors, humidity sensors, color sensors, light intensity sensors, pressure sensors, etc.


Wireless earpiece devices incorporating low-profile sensors and other electronics, according to embodiments of the present invention, offer a platform for performing near-real-time personal health and environmental monitoring in wearable, socially acceptable devices. The capability to unobtrusively monitor an individual's physiology and/or environment, combined with improved user compliance, is expected to have significant impact on future planned health and environmental exposure studies. This is especially true for those that seek to link environmental stressors with personal stress level indicators. The large scale commercial availability of this low-cost device can enable cost-effective large scale studies. The combination of monitored data with user location via GPS data can make on-going geographic studies possible, including the tracking of infection over large geographic areas. The commercial application of the proposed platform encourages individual-driven health maintenance and promotes a healthier lifestyle through proper caloric intake and exercise.


Accordingly, some embodiments of the present invention combine a personal communications earpiece device with one or more physiological and/or environmental sensor. Other embodiments may combine physiological and/or environmental sensors into an earpiece device.


Embodiments of the present invention are not limited to devices that communicate wirelessly. In some embodiments of the present invention, devices configured to monitor an individual's physiology and/or environment may be wired to a device that stores and/or processes data. In some embodiments, this information may be stored on the earpiece module itself.



FIG. 1 is a block diagram illustrating an earpiece module 100, according to some embodiments of the present invention. The illustrated earpiece module 100 includes one or more of the following: at least one physiological sensor 101, at least one environmental sensor 102 (also referred to as an external energy sensor), at least one signal processor 103, at least one transmitter/receiver 104, at least one power source 106, at least one communication & entertainment module 107, at least one earpiece attachment component 105, and at least one housing 108. Though the health and environmental sensor functionality can be obtained without the communication and entertainment module 107, having this additional module may promote use of the earpiece module 100 by users. The illustrated earpiece module 100 is intended primarily for human use; however, the earpiece module 100 may also be configured for use with other animals having ears sufficient to support an earpiece, such as primates, canines, felines, cattle, and most other mammals.


A physiological sensor 101 can be any compact sensor for monitoring the physiological functioning of the body, such as, but not limited to, sensors for monitoring: heart rate, pulse rate, breathing rate, blood flow, heartbeat signatures, cardio-pulmonary health, organ health, metabolism, electrolyte type and concentration, physical activity, caloric intake, caloric metabolism, metabolomics, physical and psychological stress levels and stress level indicators, physiological and psychological response to therapy, drug dosage and activity (drug dosimetry), physiological drug reactions, drug chemistry in the body, biochemistry, position & balance, body strain, neurological functioning, brain activity, brain waves, blood pressure, cranial pressure, hydration level, auscultatory information, auscultatory signals associated with pregnancy, physiological response to infection, skin and core body temperature, eye muscle movement, blood volume, inhaled and exhaled breath volume, physical exertion, exhaled breath physical and chemical composition, the presence, identity, and concentration of viruses & bacteria, foreign matter in the body, internal toxins, heavy metals in the body, anxiety, fertility, ovulation, sex hormones, psychological mood, sleep patterns, hunger & thirst, hormone type and concentration, cholesterol, lipids, blood panel, bone density, body fat density, muscle density, organ and body weight, reflex response, sexual arousal, mental and physical alertness, sleepiness, auscultatory information, response to external stimuli, swallowing volume, swallowing rate, sickness, voice characteristics, tone, pitch, and volume of the voice, vital signs, head tilt, allergic reactions, inflammation response, auto-immune response, mutagenic response, DNA, proteins, protein levels in the blood, body hydration, water content of the blood, pheromones, internal body sounds, digestive system functioning, cellular regeneration response, healing response, stem cell regeneration response, and the like. Vital signs can include pulse rate, breathing rate, blood pressure, pulse signature, body temperature, hydration level, skin temperature, and the like. A physiological sensor may include an impedance plethysmograph for measuring changes in volume within an organ or body (usually resulting from fluctuations in the amount of blood or air it contains). For example, the earpiece module 100 may include an impedance plethysmograph to monitor blood pressure in real-time.


An external energy sensor 102, serving primarily as an environmental sensor, can be any compact sensor for monitoring the external environment in the vicinity of the body, such as, but not limited to, sensors for monitoring: climate, humidity, temperature, pressure, barometric pressure, pollution, automobile exhaust, soot density, airborne particle density, airborne particle size, airborne particle shape, airborne particle identity, volatile organic chemicals (VOCs), hydrocarbons, polycyclic aromatic hydrocarbons (PAHs), carcinogens, toxins, electromagnetic energy (optical radiation, X-rays, gamma rays, microwave radiation, terahertz radiation, ultraviolet radiation, infrared radiation, radio waves, and the like), EMF energy, atomic energy (alpha particles, beta-particles, gamma rays, and the like), gravity, light properties (such as intensity, frequency, flicker, and phase), ozone, carbon monoxide, greenhouse gases, CO2, nitrous oxide, sulfides, airborne pollution, foreign material in the air, biological particles (viruses, bacteria, and toxins), signatures from chemical weapons, wind, air turbulence, sound and acoustical energy (both human audible and inaudible), ultrasonic energy, noise pollution, human voices, animal sounds, diseases expelled from others, the exhaled breath and breath constituents of others, toxins from others, bacteria & viruses from others, pheromones from others, industrial and transportation sounds, allergens, animal hair, pollen, exhaust from engines, vapors & fumes, fuel, signatures for mineral deposits or oil deposits, snow, rain, thermal energy, hot surfaces, hot gases, solar energy, hail, ice, vibrations, traffic, the number of people in a vicinity of the user, the number of people encountered throughout the day, other earpiece module users in the vicinity of the earpiece module user, coughing and sneezing sounds from people in the vicinity of the user, loudness and pitch from those speaking in the vicinity of the user, and the like.


In some embodiments, a physiological sensor 101 and/or an environmental sensor 102 may be configured to identify a person to whom the earpiece module 100 is attached.


In some embodiments, a physiological sensor 101 and/or an environmental sensor 102 may be configured to monitor physical aging rate of a person or subject. The signal processor 103 may be configured to processes information from a physiological sensor and/or an environmental sensor to assess aging rate. Physiological sensors configured to assess aging rate may include pulse rate sensors, blood pressure sensors, activity sensors, and psychosocial stress sensors. Environmental sensors configured to assess aging rate may include UV sensors and pollution sensors.


In some embodiments, a physiological sensor 101 and/or an environmental sensor 102 may be configured to be regenerated through a physical and/or chemical change. For example, it is anticipated that an earpiece module 100, or other device incorporating physiological and/or environmental sensors according to embodiments of the present invention may be coupled to an apparatus that is configured to “recharge” or regenerate one or more environmental and/or physiological sensors via a physical process or a chemical process, etc.


Because the earpiece module is capable of measuring and transmitting sensor information in real-time over a duration of time, the physiological and environmental sensors 101, 102 can be used to sense the aforementioned parameters over time, enabling a time-dependent analysis of the user's health and environment as well as enabling a comparison between the user's health and environment. Combined with proximity or location detection, this allows an analysis for pinpointing the location where environmental stress and physical strain took place. The signal processor 103 provides a means of converting the digital or analog signals from the sensors 101, 102 into data that can be transmitted wirelessly by the transmitter 104. The signal processor 103 may be composed of, for example, signal conditioners, amplifiers, filters, digital-to-analog and analog-to-digital converters, digital encoders, modulators, mixers, multiplexers, transistors, various switches, microprocessors, or the like. For personal communication, the signal processor 103 processes signals received by the receiver 104 into signals that can be heard or viewed by the user. The received signals may also contain protocol information for linking various telemetric modules together, and this protocol information can also be processed by the signal processor 103. The signal processor 103 may utilize one or more “compression/decompression algorithms used in digital media” (CODECs) for processing data. The transmitter 104 can be comprised of a variety of compact electromagnetic transmitters. A standard compact antenna is used in the standard Bluetooth headset protocol, but any kind of electromagnetic antenna suitable for transmitting at human-safe electromagnetic frequencies may be utilized. The receiver 104 can also be an antenna. In some embodiments, the receiving antenna and the transmitting antenna are physically the same. The receiver/transmitter 104 can be, for example, a non-line-of-sight (NLOS) optical scatter transmission system. These systems typically use short-wave (blue or UV) optical radiation or “solar blind” (deep-UV) radiation in order to promote optical scatter, but IR wavelengths can also suffice. Additionally, a sonic or ultrasonic transmitter can be used as the receiver/transmitter 104 of the earpiece module 100, but preferably using sounds that are higher or lower than the human hearing range. A variety of sonic and ultrasonic receivers and transmitters are available in the marketplace and may be utilized in accordance with embodiments of the present invention. If a telecommunication device 210 (FIG. 2) receiving wireless data signal 109 from the earpiece module 100 is in close proximity to the earpiece module, a variety of transmission schemes can be used. For communicating audible conversational information directly to the earpiece user, encoded telemetric conversational data received by the receiver 104 can be decoded by the signal processing module 103 to generate an electrical signal that can be converted into audible sound by the communication module 107.


In some embodiments, the transmitter/receiver 104 is configured to transmit signals from the signal processor to the remote terminal following a predetermined time interval. For example, the transmitter may delay transmission until a certain amount of detection time has elapsed, until a certain amount of processing time has elapsed, etc.


The power source can be any portable power source 106 capable of fitting inside the earpiece module housing. According to some embodiments, the power source 106 is a portable rechargeable lithium-polymer or zinc-air battery. Additionally, portable energy-harvesting power sources can be integrated into the earpiece module 100 and can serve as a primary or secondary power source. For example, a solar cell module can be integrated into the earpiece module 100 for collecting and storing solar energy. Additionally, piezoelectric devices or microelectromechanical systems (MEMS) can be used to collect and store energy from body movements, electromagnetic energy, and other forms of energy in the environment or from the user himself. A thermoelectric or thermovoltaic device can be used to supply some degree of power from thermal energy or temperature gradients. In some embodiments, a cranking or winding mechanism can be used to store mechanical energy for electrical conversion or to convert mechanical energy into electrical energy that can be used immediately or stored for later.


The various components describe above are configured to fit within the earpiece housing 108 and/or be attached thereto. The earpiece module housing 108 may be formed from any safe and comfortable solid material, such as metal, rubber, wood, polymers, ceramic, organic materials, or various forms of plastic. The earpiece attachment component 105 is attached to the earpiece module housing 108 and is designed to fit around or near the ear. For example, the standard Bluetooth headset includes an earpiece attachment that is connected to the headset housing via a double-jointed socket, to provide comfort and positioning flexibility for the user. In some embodiments, the earpiece attachment component 105 can be part of the housing 108, such that the entire earpiece module is one largely inflexible, rigid unit. In such case, a counterweight may be incorporated into the earpiece module 100 to balance the weight of the earpiece electronics and power source. In some embodiments, the earpiece attachment component 105 can contain physiological and environmental sensors, and the earpiece attachment component 105 may be detachable. In some embodiments, more than one earpiece attachment 105 can be attached to the earpiece module housing 108.


The communication and entertainment module 107 (also interchangeably referred to as a “communication module”) is used for, but not limited to: processing or generating an audible sound from information received via the receiver 104 (from a cell phone, computer, network, database, or the like) and/or processing or generating an electrical signal from an audible sound from the user such that the electrical signal can be transmitted telemetrically via the transmitter 104. For example, in standard Bluetooth protocol, communication electronics are used to convert an audible conversation into an electrical signal for telemetric conversation; communication electronics are also used to convert a digitized telemetric conversation into an audible conversation for the earpiece user. Additionally, the communication and entertainment module 107 can be used to store, process, or play analog or digital information from music, radio shows, videos, or other audible entertainment and to communicate this information to an earpiece user. In many cases, this information includes information received by the receiver 104. In many cases, the analog or digital information is not stored in the communication and entertainment module 107 but, rather, is stored in a portable telecommunication device 210 (FIG. 2). In such case, the communication and entertainment module 107 is used for converting the analog or digital information into audible sound for the earpiece user. The communication and entertainment module 107 may contain at least one microphone, speaker, signal processor (similar to 103), and digital memory. In some embodiments, the communication and entertainment module 107 may apply at least one CODEC for encoding or decoding information. The communication and entertainment module may utilize non-audible forms of communication with the user, such as visual, physical, or mental (i.e., brainwaves or neural stimulation) communication with the user.


In some embodiments, an audible communicator is provided that is configured to communicate therapeutic sounds (e.g., music therapy, etc.) to the person in response to physiological or psychosocial stress. The audible communicator may be embodied in the communication and entertainment module 107 or may be a separate speaker. In some embodiments, light therapy may be provided to the person in response to physiological or psychosocial stress. In some embodiments, the communication and entertainment module 107 may be configured to communicate a treatment, therapy, and/or plan of action to the person upon detection of physiological and/or environmental concerns. For example, if it is detected that the person is being exposed to unhealthy doses of UV radiation, the communication and entertainment module 107 may audibly instruct the person to move away from the person's current location (e.g., move indoors, etc.).


Like the other components of the earpiece module 100 shown in FIG. 1, the components of the communication and entertainment module 107 are not necessarily located in the same physical vicinity. The microphone and speaker of the communication module 107, for example, are located closer to the mouth and ear respectively. Furthermore, the signal processor 103 can be composed of several components located throughout the earpiece module. It should be understood that the word “module” does not necessarily imply a unified physical location. Rather, “module” is used to imply a unified function.


Bluetooth devices conventionally contain a communication module, such as communication module 107, for converting digital or analog information into audible sounds for the user. However, when combined with the health and environmental monitoring properties of an earpiece module 100 according to embodiments of the present invention, the communication and entertainment module 107 can provide functionality. For example, the earpiece module can serve as a biofeedback device. As a non-limiting example, if a user is in a polluted environment, such as air filled with VOCs, the communication module 107 may notify the user to move to a new environment. As another example, if one or more of the physiological and environmental sensors 101, 102 of the earpiece module 100 pick up a high particulate density in the environment, with an elevation in core body temperature, and a change in voice pitch occurring simultaneously (or near-simultaneously) within a common timeframe, the communication module 107 may alert the user that he/she may be having an allergic response. As a further example, the user can use the communication and entertainment module 107 to execute biofeedback for willfully controlling blood pressure, breathing rate, body temperature, pulse rate, and the like. The communication module 107 may utilize audible or visible alerts if the user is meeting their physiological targets or exceeding safe physiological limits. Alerting a user by physical or electrical force, such as the sense of touch or tingling from an electric pulse or vibration, can also be utilized. Thus, although communication by audible means is often utilized, the communication module 107 can alert, signify, or communicate with the user through sound, light, electrical actuation, and physical actuation.


As a second example of this biofeedback method, basic vital signs collected by the physiological sensors 101 and processed by the signal processor 103 can be presented to the earpiece user audibly, through the communication and entertainment module 107. For example, the user may be able to listen to his/her breathing rate, pulse rate, and the like. Additionally, an entertaining or aggravating sound or song can be used to alert the user to favorable or unfavorable personal health and environmental factors occurring in real-time. This technique may be applied towards education, such as positive or negative feedback for educational games, learning games, or games of deception (e.g., poker, etc.).


In some embodiments, the earpiece module 100 may be configured to deliver and/or monitor drugs. For example, a transdermal drug delivery system may be provided that is controlled by earpiece electronics. Earpiece sensors can monitor the drug dosage and the physiological effects of the drug in real-time.


A health and environmental network 200 according to embodiments of the present invention that may incorporate the earpiece module 100 of FIG. 1 is illustrated in FIG. 2. The earpiece module 100 is a specific sensor module 201 of the network 200, though other modules located at various other parts of the body can be used in conjunction with, or in place of, the earpiece module 100. The terms “earpiece module 100” and “sensor module 200” are used interchangeably herein in accordance with various embodiments of the present invention. The health and environmental network 200 is composed of at least one sensor module 201 (e.g., earpiece module 100) at least one portable telecommunication module 210, at least one transmission system 211, at least one user interface 214, at least one personal database 212, and at least one anonymous database 213.


The sensor module 201 can be composed of a primary module alone or a primary module and at least one secondary module. The secondary modules can be located at any location of the body, but are preferably located in a region near the ear, and preferably the earpiece module 100 serves as the primary module. In most cases, the secondary modules are not necessary. But in some cases, secondary modules may be located, for example, behind the ear (near the lymph nodes), at or near the earlobes (such as one or more earrings or ear clips), at the front of the ear (near the carotid artery), at the temples, along the neck, or other locations near the ear. These wearable secondary modules can be connected with either a “hard” connection to the primary module (such as an electric cable) or a “soft” connection to the primary module (such as a wireless connection). In Bluetooth protocol, each secondary module can be simultaneously in direct wireless communication with the primary module. Primary modules and secondary modules in the same location can promote quick-donning, ease-of-use, and comfortability for the end user. Users are not prone to accept multiple modules at multiple locations of the body.


The earpiece sensor module 201 communicates wirelessly with the portable telecommunication device 210, preferably in an open architecture configuration, such as Bluetooth or ZigBee. The telecommunication device 210 can be any portable device, such as a cell phone, PDA, laptop computer, Blackberry, another earpiece, or other portable, telemetric device. The portable telecommunication device 210 and the earpiece module 201 can telemetrically communicate both to and from each other. Though the main purpose of the portable telecommunication device is to transmit the local wireless signal from the sensor module 101 over longer distances unattainable by the transmitter 104 of the sensor module 201, the telecommunication 210 can also serve as a method of personal communication and entertainment for the earpiece user.


In some embodiments, the telecommunication device 210 transmits data in only one direction or particular directions. For example, in one embodiment, the portable telecommunication device 210 can receive telemetric information from the sensor module 201 but cannot send out signals to a transmission system 211. The portable telecommunication device 210 may also contain an end-user graphical interface, such as a user interface 214 in the network 200, such that data from the earpiece module 201 can be stored, analyzed, summarized, and displayed on the portable telecommunication device 210. For example, charts relating health and environment, as well as real-time biofeedback and the like, can be displayed on a cell phone, media player, PDA, laptop, or other device. The telecommunication device 210 may also contain physiological and environmental sensors itself, such as blood pressure, pulse rate, and pulse-oximetry, and the like. Additionally, the telecommunication device 210 can communicate with the earpiece module 201 to transfer commands, activate or deactivate sensors, communicate with the user, and the like.


The portable telecommunication device 210 sends/receives wireless information directly to/from a transmission system 211 for transmission to a database (such as personal database 312 and/or anonymous database 313) for storage, analysis, and retrieval of data. The style of transmission system depends largely on the location of the database. For example, if the database is located in a local computer, the wireless information from the telecommunication device 210 can be sent directly to the local computer. This computer may be connected with the Internet, allowing access to the database from the web. However, the database is more typically located far away from the user and telecommunication module. In this case, the wireless signal from the telecommunication device 210 can be sent to a reception tower and routed through a base station. This information can then be sent to a database through the Internet. A variety of other transmission protocols can be applied for connection between the telecommunication device 210 and the databases 212, 213.


The personal and anonymous databases 212, 213 represent databases that may or may not be located on the same computer. A key difference between these two databases is not the physical location of the database but rather the type of information available on each database. For example, the anonymous database 213, containing health and environmental data from multiple indistinct earpiece users, can be public and accessible through the Internet by various users. In contrast, the personal database 212 contains health and environmental data that is personalized for each user, including personalized information such as name, birth date, address, and the like. Users can log-in to their personalized information in the personal database 212 through an interactive user interface 214 and compare this information with information from multiple users in the anonymous database 213 via a graphical user interface.


The user interface 214 can be a computer monitor, a cell phone monitor, a PDA monitor, a television, a projection monitor, a visual monitor on the earpiece module 201, or any method of visual display. (Audible methods and audio-visual methods can also be used for the user interface 214.) For example, the user may log-in to their personal database 212 through a computer user interface 214 and compare real-time personal health and environmental exposure data with that of other users on the network 200. In some cases, the data from other users may be anonymous statistics. In some cases, one or more users may have agreements to view the data of one or more other users, and in other cases, users may agree to share mutual personalized data through the Internet.


A specific embodiment of a graphical user interface 300 is presented in FIG. 3. FIG. 3 shows an example of how a computer monitor may appear to a user logging-in to their personal database 212 and comparing their own personal data with that of anonymous users in the same network 200. In this case, data from anonymous users is averaged into certain demographics; the choice of the demographics to be displayed can be selected by the user accessing the personalized database. In the graphical user interface 300 of FIG. 3, the user's personalized data, signified by a star, is compared with statistics from other users in an anonymous database 213. This allows the user to compare his/her health and environment with that of others in selected demographics. Thus, this network 200 serves not only as a source of useful information from a medical standpoint, but also as a form of entertainment for curious users.


The network 200 can be used in medicine for a variety of important functions. As one example, a doctor can monitor the health of patients through each patient's personalized database 212. If the earpiece module 201 contains a dosimeter, the doctor can even monitor the efficacy of prescribed medications, and the physiological response to medications, over time. This dosimetry approach is directly applicable to clinical studies of various treatments. For example, during a clinical trial, the earpiece module 201 can collect environmental data, drug dosimetry data, and physiological data from the earpiece user such that researchers can understand the etymology between drugs, genes, physiology, environment, and personal health.


Because of the high compliance of the earpiece module 100, primarily due to the dual-modality as a health/environmental monitor and a personal communication/entertainment device, users are prone to wear this device throughout clinical trials, providing more valuable information for drug discovery and the pharmaceuticals market.


As a further example, the health and environmental network 200 can be used by dieticians to track the caloric intake, health, and physical activity of dieters. Similarly, the network 200 can be used by athletic trainers to monitor the diet, physical activity, health, and environment of athletes. In many cases professionals are not necessary, and the user can monitor his/her own diet, activity, athletic performance, etc. through the network without professionals, parents, guardians, or friends monitoring their personal statistics.


In a specific example of the network 200, an earpiece user is a test subject in a clinical trial for a new treatment, such as a new drug, physical therapy, medical device, or the like. The earpiece user's health and environment are monitored in real-time, and this data is stored on the earpiece module 201, the portable telecommunication device 210, the personal database 212, or the anonymous database 213. By accessing the stored data, researchers managing the clinical trial can then compare the statistics from multiple users to make correlations between user environment, health, and the effectiveness of treatment.


It should be noted that algorithms for processing personal health and environmental data, diagnosing medical conditions, assessing health states, and the like do not need to be limited to the illustrated network 200. Various algorithms can also be integrated into the earpiece module 201 or telecommunication device 210 according to embodiments of the present invention. A data storage component in at least one of these units allows processed signal data to be stored, analyzed, and manipulated to provide new knowledge to the user. This storage component can be any solid-state storage device, such as flash memory, random-access memory (RAM), magnetic storage, or the like. For example, the earpiece module 201 can be programmed to monitor certain habits, such as nail-biting. In this non-limiting example, the earpiece module physiological sensors 101 may monitor internal sounds, and an algorithm can be implemented to monitor signatures of nail-biting sounds in real-time. If the habit is identified by the algorithm, the earpiece communication module 107 may instantly warn the user that the habit is occurring. Alternatively, the algorithm may count the number of times a day the habit occurred, monitor physiological and psychological stress indicators during each occurrence, log each time when the habit occurred, and store environmental data associated with the habit. This stored data can be accessed at a later time, allowing the user to determine what environmental factors cause the physiological or psychological stress associated with nail-biting. As this example shows, these algorithms can take advantage of both physiological sensor 101 data and environmental sensor 102 data.


A data storage component may include various algorithms, without limitation. In some embodiments, at least one algorithm is configured to focus processing resources on the extraction of physiological and/or environmental information from the various environmental and/or physiological sensors. Algorithms may be modified and/or uploaded wirelessly via a transmitter (e.g., receiver/transmitter 104 of the earpiece module 100)


The biofeedback functionality of the telemetric earpiece module 100 can be applied towards various gaming applications. For example, one or more earpiece users can connect their earpiece modules 100 to one or more gaming devices wirelessly through the open architecture network provided by Bluetooth, ZigBee, or other such networks. This allows personal health and environmental information to be transferred wirelessly between the earpiece module 100 and a gaming device. As earpiece users play a game, various personal health and environmental feedback can be an active component of the game. In a non-limiting embodiment, two users playing a dancing game, such as Dance Dance Revolution, can monitor their vital signs while competing in a dancing competition. In some cases, users having healthier vital signs, showing improved athletic performance, will get extra points (“Vital Points”). In another specific example, this personal health and environmental information can be sent telemetrically to a gaming device to make entertaining predictions about one or more users. Namely, the gaming device may predict someone's life expectancy, love-life, future occupation, capacity for wealth, and the like. These predictions can be true predictions, purely entertaining predictions, or a mixture of both. Sensors measuring external stressors (such as outside noise, lighting conditions, ozone levels, etc.) and sensors measuring internal stresses (such as muscle tension, breathing rate, pulse rate, etc.) integrated into the earpiece module 100 can be used to facilitate predictions by the gaming device. For example, the information from the sensors can be recorded from one or more earpiece users during a series of questions or tasks, and the information can be sent telemetrically to a gaming device. An algorithm processed in the gaming device can then generate an entertaining assessment from the information. This game can be in the form of a video game, with a graphical user interface 214, or it can be a game “in person” through an entertainer. Other games can involve competitions between multiple earpiece monitor users for health-related purposes, such as online dieting competitions, fitness competitions, activity competitions, or the like. Combining the telemetric earpiece module 100 with gaming, according to embodiments of the present invention, provides seamless interaction between health and environmental monitoring and the game, through a comfortable telemetric module. Other sensor modules 201 located at other parts of the body can also be used.


An additional non-limiting embodiment of the biofeedback functionality of the earpiece module 201 can be monitoring psychological and physiological stress (such as monitoring stress indicators) during a poker game. These stress indicators can be breathing rate, muscle tension, neurological activity, brain wave intensity and activity, core body temperature, pulse rate, blood pressure, galvanometric response, and the like. Users may, for example, use the earpiece module 201 to record or display their psychological and physiological stress during a poker game in real-time. This information can be stored or displayed on a portable telecommunication device 210 or sent wirelessly to other parts of the network 200. The user can use this biofeedback to adjust their psychological and physiological stress (or stress indicators) through force of will. This biofeedback process allows earpiece users to self-train themselves to project a certain “poker face,” such as a stoic cold look, a calm cool look, or another preferred look. Additionally, external stressors, such as light intensity and color, external sounds, and ambient temperature, can be sensed, digitized, and transmitted by the earpiece module 100 to a telecommunication device (for storage), providing the user with important information about how the external environment may be affecting their stress response and, hence, poker game. In some games, the stress indicators may be displayed for outside viewers (who are not part of the poker game) as a form of entertainment when watching a group of poker players (each having earpiece modules 201) in a casino, television, or through the Internet.


The biofeedback approach is also directly relevant to personal education as a learning tool. For example, monitoring the physiological and psychological response to learning can be used to help users understand if they are learning efficiently. For example, in the course of reading, the earpiece module 201 can monitor alertness through galvanometric, brainwave, or vital sign monitoring. The user can then use this information to understand what reading methods or materials are stimulating and which are not stimulating to the earpiece user.


In the broader sense, the discussed earpiece-enabled biofeedback method can be used as a self-training tool for improving performance in public speaking, athletic activity, teaching, and other personal and job-related activities.


The health and environmental network 200 enables a variety of additional business methods. For example, users can be charged a fee for downloading or viewing data from the personal and/or anonymous databases 212, 213. Alternatively, users may be allowed free access but be required to register online, providing personal information with no restrictions on use, for the right to view information from the databases. In turn, this personal information can be traded or sold by the database owner(s). This information can provide valuable marketing information for various companies and government interests. The health and environmental data from the databases 212, 213 can be of great value itself, and this data can be traded or sold to others, such as marketing groups, manufacturers, service providers, government organizations, and the like. The web-page or web-pages associated with the personal and anonymous databases 212, 213 may be subject to targeted advertising. For example, if a user shows a pattern of high blood pressure on a personal database 212, a company may target blood pressure treatment advertisements on the user interface 214 (i.e. web page) while the user is logged-in to the personal database through the user interface 214. For example, because various health and environmental statistics of earpiece users will be available on the database, this information can be used to provide a targeted advertising platform for various manufacturers. In this case, a database manager can sell information to others for targeted advertising linked to a user's personal statistics. In some cases, a database owner does not need to sell the statistics in order to sell the targeted advertising medium. As a specific example, a company can provide a database owner with statistics of interest for targeted advertising. For example, the company may request advertising a weight-loss drug to anonymous users having a poor diet, high caloric intake, and/or increasing weight. A database manager can then charge the company a fee for distributing these advertisements to the targeted users as they are logged-in to the database website(s). In this way, the users remain anonymous to the company. Because targeted advertisements can be such a lucrative market, income from these sources may eliminate the need for charging users a fee for logging-in to the databases 212, 213.


The earpiece module 201 and earpiece module network 200 can enable a variety of research techniques. For example, a plurality of earpiece modules 100 worn by users can be used in marketing research to study the physiological and psychological response of test subjects to various marketing techniques. This technique solves a major problem in marketing research: deciphering objective responses in the midst of human subjectivity. This is because the physiological and psychological response of the earpiece user largely represents objective, unfiltered information. For example, users that are entertained by a pilot TV program would have difficulty hiding innate vital signs in response to the program. The data generated by the earpiece module 201 during market research can be transmitted through any component of the telemetric network 200 and used by marketing researchers to improve a product, service, or method.


Another business method provided by the network 200 is to charge users of the network for usage and service (such as compilation service). For example, a clinical trial company may pay a fee for accessing the databases 212, 213 of their test subjects during medical research. In this case, these companies may buy earpiece modules 201 and pay for the service, or the earpiece modules 201 may be provided free to these companies, as the database service can provide a suitable income itself. Similarly, doctors may pay for this service to monitor patients, fire fighters and first responders may pay for this service to monitor personnel in hazardous environments, and athletic trainers may pay for this service to monitor athletes. Also, users can pay for the database service directly themselves. Because these databases 212, 213 are dynamic, updated regularly via the earpiece module 201 of each user, and changing with time for individual users and users en mass, these databases can maintain a long-term value. In other words, there may always be new information on the databases 212, 213.


Another embodiment of the present invention involves methods of combining information from various earpiece health sensors into a meaningful real-time personal health and environmental exposure assessment in a recording device. The meaningful assessment is generated by algorithms that can be executed in the earpiece 201, in the portable telecommunication device 210, or through various other electronic devices and media within the network 200. In one embodiment, raw or preprocessed data from the sensor module 201 is transmitted wirelessly to the telecommunication device 210, and this device executes various algorithms to convert the raw sensor data (from one or more sensors) into a meaningful assessment for the user. In another embodiment these algorithms are executed within the earpiece 201 itself, without the need for processing in the telecommunication device 210. The output from these algorithms can be viewed as charts, graphs, figures, photos, or other formats for the user to view and analyze. Preferably, these formats display various health factors over time with respect to a particular environment, with health factor intensity on the dependent axis and time or environmental factor intensity on the independent axis. However, virtually any relationship between the physiological data and environmental data can be processed by the algorithm, and these relationships can be quantitative, qualitative, or a combination of both.


One innovation involves applying the earpiece sensor module 201 towards a physical or mental health assessment method. An algorithm may combine data from health and environmental sensors 101, 102 towards generating a personal overall health assessment for the user, conditional to a particular environment. For example breathing rate, pulse rate, and core body temperature can be compared with ozone density in the air for generating an ozone-dependent personal health assessment. In another specific example of this innovation, information from the earpiece sensors 101, 102 can be used to monitor overall “mood” of a user in a particular environment. More particularly, algorithmic processing and analyzing of data from sensors for core body temperature, heart rate, physical activity, and lighting condition can provide a personal assessment of overall mood conditional on external lighting conditions.


As previously mentioned, the ear is located at an ideal physiological position for monitoring a variety of health and environmental factors. Thus, the ear location can enable a variety of methodologies for physiological and environmental monitoring with an earpiece module 100. In particular, because the ear canal is naturally designed for the transmission of audible sound, the ear canal facilitates methods for monitoring physiological processes by monitoring internal sounds. However, when extracting physiological information from the body, in a given external environment, environmental information is inevitably part of the extracted signal. This is because external energy, such as external audible noise, is entering the body. Thus, when listening to internal sounds, external sounds are also picked up. A methodology for cleaning up the signal such that it contains clearer information about physiology (as opposed to external environment) is provided by some embodiments of the present invention.



FIG. 4 illustrates a physiological signal extraction methodology 400 for selectively monitoring internal physiological energies through an earpiece module 420 according to embodiments of the present invention. In the illustrated method, internal physiological energy is sensed by a sensor designated “Sensor-1” 401. Sensor-1 generates an electrical signal in response to the physiological energy. One or more external sensors 402, 403, and 404 sense external energy from the environment in the vicinity of the earpiece module user and generate an electrical signal in response to the external energy. Though only one external sensor is needed, multiple sensors can be used to add sensing functionality, improve signal extraction, and/or increase the selectivity of sensing various energies. In FIG. 4, the external energy sensors 402, 403, 404 are shown collecting energy from different directions to emphasize that each sensor can be sensing the same type of energy but from a different direction, as this directional information can be useful for various assessments of the earpiece user. The energies described can be any physical energy, such as electrical, magnetic, electromagnetic, atomic, gravity, mechanical, acoustic, and the like. A signal processor 405 collects the electrical sensor responses and processes these signals into a signal that can be transmitted wirelessly through a transmitter/receiver 406 for communicating the information 407 telemetrically between the earpiece module 420 and a portable telecommunication device 210 (FIG. 2).


As with processor 103 of FIG. 1, the signal processor 405 of FIG. 4 can be used to combine signals from the various sensors, compare similarities between the signals, and generate a new signal that contains cleaner physiological information than any of the original signals. This can be done by converting the analog signals from the physiological sensors 401 and environmental sensors 402, 403, and 404 into digital signals and comparing the signals in, for example, a digital comparator to form a new signal that contains cleaner physiological information. (In some cases, an analog comparator technique can also be used if the signals are not digitized.) If these digitized signals are synchronized in time, a subtraction of environmental features from the signals can be realized by the comparator. Further, if algorithms are integrated into the signal processor 405, comparisons can be made with respect to how external energy affects physiological energy in time.


An embodiment of the physiological signal extraction methodology 400 is presented in FIG. 5 as an acoustical-cancellation physiological signal extraction methodology 500. An earpiece module 517 (for example, with the functionality of earpiece module 100 of FIG. 1) is attached to the ear 506 with an ear attachment 505. This earpiece module 517 is physically similar, if not identical, to the various examples shown in FIGS. 9-16, discussed below. The earpiece module 517 contains at least one physiological acoustical sensor 501 pointed in the direction of the tympanic membrane 520 and at least one external acoustical sensor 502 pointing away from the body and towards the outside environment. To suppress the convolution of external and internal sounds, an acoustical buffer region 519 is placed between the two sensors 501, 502. Environmental sounds 507 are sampled by the external acoustical sensor 502, and physiological sounds 508 traveling through the ear canal 530 (and towards the earpiece 517) are sampled by the physiological acoustical sensor 501. Because the tympanic membrane and other body parts and tissues naturally vibrate in response to external sounds 507, part of the physiological acoustical energy 508 is composed of environmental acoustical energy 507 as well as physiological sounds. These physiological sounds are referred to as “auscultatory” information. By comparing digitized signals from each acoustical sensor 501, 502, the external energy 507 signatures can be at least partially removed from the auscultatory 508 signatures such that a new signal, containing cleaner physiological information, can be generated. For example, the sounds of external steps and human voices can be digitally removed or reduced from the final processed signal such that the final signal contains a cleaner representation of the internal sounds of pulse rate, breathing rate, swallowing rate, and other auscultatory information.


The acoustical sensors 501, 502 can contain any acoustical transducer, such as a microphone, piezoelectric crystal, vibrating membrane, magnetic recorder, and the like. Further, the acoustical sensors 501, 502 may contain a variety of layers for filtering sounds and promoting the directional extraction of sound. Additionally, various electrical filters, such as low-pass, high-pass, band-pass, notch, and other filters, can be used to clean-up signals from each sensor 501, 502 to help remove unwanted sounds or signatures. In some embodiments, the acoustical sensors 501, 502 are compact microphones, such as electric microphones or piezoelectric microphones, and the signals from these microphones are electrically filtered. The acoustical buffer region 519 can be any material that absorbs acoustical energy. In some embodiments, this material is soft, durable material, such as plastic, foam, polymer, or the like. In some embodiments, the acoustical buffer region 519 can be specially shaped to reflect or absorb sounds of certain frequencies through acoustical interference.


An example of how the auscultatory signal extraction technique 500 may be used is summarized in test data 800 set forth in FIG. 8. A small microphone was placed inside the ear of a test subject, and various sounds were recorded over time. In the test data 800, the user was relaxing on a chair. The raw waveform 801 contains information from internal and external sounds. However, following digital filtering and noise cancellation, the final energy spectrum 802 and waveform 803 contain cleaner information about the test subject's pulse rate. In fact, the signature of each pulse can even be identified. In processing this signal, the noise reduction algorithm was selected by intelligently choosing a data segment dominated by external (environmental) acoustical energy, where physiological information was largely not present. Such intelligent algorithms can be integrated into a Bluetooth earpiece module for automatic auscultatory analysis for extracting physiological sounds from the body. Because the nature of external sounds and internal sounds are known by the placement of sensors 501, 502, providing a basis for signal subtraction, the acoustical signal extraction innovation 500 can provide clean auscultatory data automatically in real-time.


Another embodiment of the signal extraction methodology 400 is presented in FIG. 6. The optical physiological signal extraction technique 600 is a method of extracting a variety of physiological information form the tympanic membrane 620 by locking-in to the vibrational frequency of the tympanic membrane 620. In this method 600, the earpiece module 614 contains at least one acoustical energy sensor 601 for measuring acoustical energy coming from the ear canal 630 and other neighboring organs and tissues. At least one optional external acoustical energy sensor 602 can be used for measuring environmental sounds in the vicinity of the earpiece user, as the combined signals from sensors 601, 602 can produce a cleaner signal for physiological monitoring. At least one optical emitter 624 is located in the earpiece for generating optical energy directed towards the tympanic membrane 620 through the ear canal 630. This optical energy is absorbed, scattered, and reflected by the ear canal 630 and tympanic membrane 620. In some cases, the optical energy induces fluorescence in the ear canal or tympanic membrane. In other cases, the optical energy experiences a change in polarization or other optical properties. In many cases, a change in more than one optical property (absorption, reflection, diffraction, fluorescence, polarization, etc.) occurs. In any case, this resulting optical energy is referred to as the “secondary energy.” The secondary energy is detected by at least one optical detector 625, though more than one optical detector 625 may be utilized. As with the acoustical sensors 601, 602, the optical detector 625 converts incoming energy (in this case optical energy) into an electrical signal to be sent to a signal processor 405 (FIG. 4). The optical detector 625 may contain filters for selectively passing optical energy of physiological interest. A buffer region 619 is used to prevent external sound and light from convoluting the extraction of physiological information. In many cases, the optical energy generates a secondary response that is not optical in origin, such as a thermal response or biochemical response. In such case, at least one optical detector 625 may be replaced with at least one other type of sensor for sensing the non-optical secondary response.


Because the thin tympanic membrane 620 vibrates significantly in response to sound, whereas the other physiological ear features, such as the ear canal 630 and external ear 606 do not vibrate significantly, a method of extracting secondary optical signals scattered from the vibrating tympanic membrane 620 is provided. Acoustical information from the tympanic membrane vibrational response, collected by the acoustical sensor 601 (or a combination of sensors 601, 602), is processed by a signal processor 405, and the secondary optical information from the tympanic membrane is collected by the optical detector(s) 625. The signal processor compares digitized signals from these sensors in synchronized time, such that signals from the optical detector 625 containing frequency components characteristic of the tympanic membrane's vibrational response are selectively extracted to provide cleaner physiological information from the tympanic membrane. For example, the oxygen content of blood in the tympanic membrane can be monitored by the reflection of red and infrared light from the tympanic membrane, similar to pulse-oximetry. However, scattered optical energy from the ear canal may make it difficult to extract blood oxygen from the tympanic membrane, as the source of scattered light is unclear. The optical physiological signal extraction methodology 600 provides a means of locking-in to the tympanic membrane optical reflection signal through the tympanic membrane vibrational signal collected by the acoustical sensor 601 (or the combination of sensors 601, 602). The illustrated methodology 600 works because the thin tympanic membrane 620, in response to sound, vibrates with substantially greater deflection than the ear canal, and thus primary and secondary light scattered from the tympanic membrane 620 is largely modulated by the frequency of sound reaching the tympanic membrane. This same technique 600 can be applied towards monitoring scattered optical energy from the vibrating bones of the ear, using electromagnetic wavelengths capable of passing through the tympanic membrane.


Another non-limiting embodiment of the optical physiological signal extraction methodology 600 of the present invention involves dosimetry. For example, the concentration of a drug and/or the performance of a drug can be monitored over time by characterizing the real-time fluorescence response of the drug, or intentional fluophores placed in the drug, via the tympanic membrane. In such case, the fluorescence response of the tympanic membrane can be extracted from optical noise through the illustrated methodology 600. However, in this case the fluorescence response, as opposed to the reflectance response, is extracted from the tympanic membrane 620.


In the optical physiological signal extraction methodology 600, a variety of devices can be used for the optical source or optical sources 624, such as a light-emitting diode (LED), a laser diode (LD), a miniature lamp (such as a miniature incandescent lamp, a miniature mercury lamp, or a miniature tungsten lamp), a light guide delivering light from an outside source (such as the sun or other light source), a multi-wavelength source, a microplasma source, an arc source, a combination of these sources, and the like. Special variants of light-emitting diodes, such as resonant-cavity light emitting diodes (RCLEDs), superluminescent LEDs (SLEDs), organic LEDs (OLEDs), and the like can also be applied.


In some embodiments of FIG. 6, coherent light can be used to monitor physiological processes. Monitoring vibrating membranes can be accomplished by using lasers and LDs such that coherent optical energy can directly interrogate the membrane and interference signals can be extracted. However, the signal extraction approach of FIG. 6 allows physiological signal extraction from the tympanic membrane with incoherent light-emitting diodes, which use less power, are more commercially available, and are more cost-effective than laser diodes. In fact, the scattered light from LEDs can be an advantage as scattered light may be necessary for reaching the tympanic membrane from outside of the ear canal. More specifically, earpiece module users may prefer to not have a tube stuck deep into the ear canal, and thus there may be no direct, unobstructed optical path to the tympanic membrane from outside of the ear.


It should be noted that in some cases the optical physiological signal extraction methodology 600 can be implemented without the optical emitter 624. For example, the native IR blackbody radiation of the tympanic membrane, scattered in proportion to acoustical vibrational motion of the tympanic membrane, can be extracted using the approach 600 of FIG. 6 without the need of an optical emitter 624. This may be accomplished by locking-in to the acoustically modulated blackbody IR radiation from the tympanic membrane, sensed by the photodetector 625, through the signals received by the acoustic sensor 601 or 602. The extracted blackbody radiation from the tympanic membrane can then be processed by a signal processor 103 to yield a resulting signal indicative of core body temperature. In this embodiment, the photodetector 625 may be, for example, an IR sensor, an IR photodiode, an IR avalanche photodiode, an IR photoconductor, an IR-detecting field-effect transistor, a fast-response thermal sensor such as a pyroelectric sensor, or the like.


A specific pictorial example of the innovative optical physiological signal extraction methodology 600, incorporating an LED-photodetector module 700, is shown in FIG. 7. In the illustrated module 700, at least one LED 724 generates at least one optical beam 713 directed towards the tympanic membrane 720 through the ear canal 730. At least one photodetector 725 is positioned to receive scattered light 717 modulated by the tympanic membrane 720. Inevitably, scattered light from the ear canal 730 not associated with the tympanic membrane will also reach the photodetector 725. The optical physiological signal extraction methodology 600, according to some embodiments of the present invention, can be used to reduce the impact of scattered light from the ear canal 730 and increase the impact of scattered light from the tympanic membrane 720.


An optical reflector 727 may be used to steer the light from the LED 724 towards the tympanic membrane 720 and away from the photodetector 725, preventing convolution and saturation by the optical source light 713. The LED 724 and photodetector 725 are mounted onto a mounting board 726 in a discrete module, and this discrete module may be mounted to a larger board 710 for integration with circuitry in an earpiece sensor module 100. Mounting of components to the board 726 and the board 726 to the larger board 710 can be accomplished, for example, by heating soldering bumps underneath the parts through standard electronic soldering techniques. The photodetector 724 can be any solid state device, such as a photodiode, an avalanche photodiode, a photoconductor, a photovoltaic, a photomultiplier, a FET photodetector, a photomultiplier tube, or the like. In some cases, the larger mounting board 710 may be connected to a detachable element, such as a cable, jack, fixture, or the like.


The active optical absorption region of the photodetector 725 may be covered by at least one optical filter for selectively passing light of interest. Light-guiding optics may also be integrated. Optical filters and light-guiding optics may also be applied to the LED source 724. The LED 724 can be any optical wavelength from the deep-UV to the deep-IR. In some cases, the LED 724 can be replaced with a laser diode or other compact laser source, as long as electrical powering requirements are satisfied. In such case, the laser diode may need to be pulsed on a set interval to prevent a battery drain from continuous laser diode usage.


Referring back to FIG. 4, the directional external energy sensors 402, 403, and 404 can be useful for monitoring multiple sounds at once and deconvoluting interference from other sounds. For example, to monitor footsteps (pedometry), physical activity, and/or the impact on vital signs, signals from the sensors 401, 402, 403, and 404 can all be processed together, via the signal processor 405, to generate meaningful information about each factor. More specifically, the sound of footsteps can be extracted from the final processed signal by deconvoluting directional sounds from above the earpiece user, through Sensor-3 403, and by deconvoluting directional sounds from the side of the earpiece user, through Sensor-2 402. In this manner, sounds from footsteps, coming primarily from below the user, measured through Sensor-4 404, and from inside the user, measured through Sensor-1 401, can be extracted from interfering sounds coming from other, non-relevant directions.


According to some embodiments of the present invention, a person's vitals signs can be extracted through the same methodology, but in this case, the sounds measured from at least one external energy sensor (Sensor-2, Sensor-3, or Sensor-4 404) are also deconvoluted from the final signal such that the final signal contains cleaner physiological information than that from Sensor-1 401 alone. As a further example of the acoustical signal extraction methodology 500 of the present invention, the signal extraction technique can be used to extract acoustical signals associated with one or more of the following: yawning, swallowing, eating, masticating, sleeping, slurping, walking, running, physical activity, jogging, jumping, teeth grinding, jaw movements, a change in bite, changes in speech, changes in voice (volume, pitch, speed, inflammation of vocal chords, etc.), coughing, snoring, sneezing, laughing, eye muscle movements, crying, yelling, vocal stress, physical and psychological stress, stuttering, digestion, organ functioning, vital signs, pulse rate, breathing rate, cardiovascular performance, pulmonary performance, lung capacity, breathing volume, blood pressure, athletic performance, physiological or psychological stress indicators, the number of typed words on a keyboard or typing rate, personal habits (such as scratching, nail biting, saying “um,” hair pulling, smoking, and the like), emotional states, muscle tension, and the like.


It should be clear that the general physiological signal extraction methodology 500 is also applicable in the reverse. Namely, the external environmental energy can be extracted from the convolution of external energy with physiological energy through the same basic process. In such case, the signal processor 405 subtracts signatures associated with internal physiological energy such that the new processed signal contains cleaner information about the environment. It should also be clear that any of the sensors 401, 402, 403, and 404 can be composed of multiple sensors measuring multiple forms and expressions of various physical energies.


The earpiece modules described herein need not be embodied within headsets. For example, an earpiece module 100 according to embodiments of the present invention can be a hearing aid, an earplug, an entertaining speaker, the earpiece for an IPOD, Walkman, or other entertainment unit, a commercial headset for a phone operator, an earring, a gaming interface, or the like. The earpiece module 100 covers the broad realm of earpieces, ear jewelry, and ear apparatuses used by persons for entertainment, hearing, or other purposes both inside and outside of health and environmental monitoring.


Moreover, two earpiece modules 100 may be utilized, one for each ear of a person, according to some embodiments of the present invention. Dual-ear analysis with two earpiece modules can be used, for example, to compare the core temperature of each tympanic membrane in order to gauge brain activity comparing each brain hemisphere. In another case, acoustical energy, including ultrasonic energy, can be passed from one earpiece module to the other, with acoustic absorption and reflection being used to gauge various physiological states. For example, this technique can be used to gauge hydration level in the head or brain by estimating the acoustical energy absorption rate and sound velocity through the head of the user.


A variety of earpiece styles, shapes, and architectures can be used for earpiece module 100 according to embodiments of the present invention. A non-limiting embodiment of the earpiece module is shown pictorially in FIG. 9. The illustrated earpiece 905 fits over the ear of a person and is held in place by an ear support 901 (also called the “earpiece attachment component” 105). The illustrated earpiece module 905 also includes an earpiece body 902, an earpiece fitting 908, and an optional earlobe clip 904. The earpiece may also contain an adjustable mouthpiece 1416 (FIG. 14B) and/or a pinna cover 1402 (FIGS. 14A-14B) described below. The earpiece 905 connects with the ear canal of a person through an earpiece fitting 908 located on the backside 906 of the earpiece 905. The earpiece fitting 908 transmits sound to the inner ear and eardrum. Health and environmental sensors are integrated primarily within or along the earpiece body 902, including the earpiece backside 906. However, an earlobe clip 904 can contain various health and environmental sensors as well. In some cases, health and environmental sensors can be integrated within or along the ear support 901, the adjustable mouthpiece 1416, the earpiece fitting 908, or the pinna cover 1402. Raw or processed data 903 from these sensors can be wirelessly transferred to a recording device or a portable telecommunication device 210 (FIG. 2). In some embodiments of the present invention, a recording device can be located within or about the earpiece 905 itself. In some cases, this recording device is flash memory or other digitized memory storage. The types of health and environmental factors which may be monitored have been previously described above for the earpiece module 100.


It should be understood that the earpiece body 902 can be any shape and size suitable for supporting an earpiece fitting 1008. In some cases, the earpiece body and earpiece fitting can be one and the same structure, such that the earpiece body-fitting is a small fitting inside the ear. In many cases, it is desirable to seal off or partially seal off the ear canal so as to prevent sounds from entering or leaving the ear such that auscultatory signal can more easily be extracted from the ear canal through devices (such as microphones) in the earpiece body-fitting.


It should be noted that the invention is not limited to the exemplary earpiece 905 of FIG. 9. Other earpiece configurations are also capable of integrating health and environmental sensors for portable, noninvasive, real-time health monitoring according to embodiments of the present invention. For example, the earlobe clip 904 can be modified to reach other surfaces along or near a person's ear, head, neck, or face to accommodate electrical or optical sensing. Similarly, more than one clip may be integrated into the earpiece 905. Sensors can be integrated into the earpiece-fitting 1008 as shown in the earpiece 1002 of FIG. 10. In such embodiments, the sensors may be integrated into a module 1009 in the earpiece-fitting 1008. Environmental sensors are preferably located on the outside of the earpiece 1205 through a region on the earpiece frontside 1206 (as shown in FIG. 12). This allows access to air in the vicinity of the earpiece user. However, environmental sensors can be located anywhere along the earpiece module 905.



FIG. 12 illustrates details about the location of sensors in certain parts of an earpiece module 1205, according to embodiments of the present invention. The ear support 1201 contains a pinna cover 1202 that may contain sensors for monitoring physiological and environmental factors. This structure is particularly useful for sensing methodologies which require energy to be transmitted through the thin layers of the pinna (the outer ear). Though any portion of the pinna can be covered and/or contacted, in some embodiments, the pinna cover 1202 overlaps at least a part of the helix or a part of the scapha of an ear. Likewise, an optical absorption detector, composed of an optical emitter and optical detector, can be integrated into the pinna cover 1202 for monitoring, for example, hydration, dosimetry, skin temperature, inductive galvanometry, conductive galvanometry, and the like.


Galvanometry, the measurement of electrical properties of the skin, can be measured inductively, through contactless electromagnetic induction without contacts, or conductively, with two, three, four, or more conductivity probes. Additionally, a 4-point conductivity probe technique, such as that used for measuring the conductivity of semiconductor wafers, can be applied. A variety of sensors can be integrated into the earpiece fitting 1208. For example, a galvanometric device can be integrated into the surface 1209 of the earpiece fitting where the earpiece fitting touches the skin of the outer ear. Additionally, an inductive device, such as an inductive coil 1214, can be integrated along the earpiece fitting body to measure movements of the tympanic membrane inductively. The inductive impedance can also be measured with the inductive coil 1214 or another inductive sensor, and this can be applied towards contactless galvanometry. The inductive coil 1214 can include one or more coils arranged in any orientation, and a core material, such as an iron-containing material, may be used to improve the sensitivity of the response. In some cases, multiple coils may be used to facilitate the canceling of stray electromagnetic interference. Sensors can also be integrated into the end tip 1212 of the earpiece fitting 1208 to measure physiological properties deeper into the ear canal. For example, the optical module 700 of FIG. 7 may be located in, at, or near the end tip region 1212 in a module 1213. The sensors on the module 1213 in this region are carefully arranged so as not to prevent the transmission of sound (from the built-in communication module 107) and to not be distorted during earpiece placement and removal. The end tip sensor module 1213 can contain several types of sensors for generating multiple types of energy and detecting multiple types of energy, and this module can be integrated into the speaker module (part of the communication module 107) inside the earpiece fitting 1208 that is used for sound transmission to the user during telemetric conversations. In some cases, the speaker module can be used as a microphone to measure auscultatory signals from the body. This may be especially useful for measuring low frequency signals less than 1000 Hz. Employing the speaker as a microphone may require impedance matching to maximize the auscultatory signal extraction.


Alignment, placement, and arrangement of sensors, according to embodiments of the present invention, can be enabled or simplified by adopting a flexible circuitry configuration 1300, such as that shown in FIG. 13. A flexible circuit board 1304 according to embodiments of the present invention can be made out of any stable flexible material, such as kapton, polymers, flexible ceramics, flexible glasses, rubber, and the like. A key requirement of the flexible material of the flexible circuit board 1304 is that it must be sufficiently electrically insulating and electrochemically inert. As with a standard rigid circuit board, a variety of sensors 1306 can be mounted on the flexible circuit board 1304, and this board can be integrated into any part of the earpiece module 905 (FIG. 9). Flexible circuitry can be especially useful for odd-shaped components of the earpiece, such as the earpiece fitting 908, ear support 901, the earpiece clip 904, the adjustable mouthpiece 1416, and the pinna cover 1202/1402. In some cases, flexible piezoelectric polymers, such as polyvinylidene fluoride may be useful for measuring body motion and auscultatory sounds from the body.



FIGS. 14A-14B illustrate an embodiment 1400 of an earpiece module 1405 with an adjustable mouthpiece 1416 and a pinna cover 1402. The earpiece 1400 contains a region where an adjustable mouthpiece 1416 can be swiveled, extended, pulled, extracted, flipped, or ejected towards the mouth. A microphone at the end of the mouthpiece 1416 can be used to improve personal communication through the earpiece 1400. Sensors integrated into the mouthpiece 1416 can be used to monitor, for example, exhaled breath for respirometry and inhalation/exhalation monitoring. Carbon dioxide, oxygen, nitrogen, water vapor, and other respired gases and vapors can be monitored, providing an overall assessment of health. Additionally, VOC's and other vapors exhaled by the breath can be monitored for diagnosing various disease states, such as diabetes, obesity, diet, metabolism, cancer, hepatic or renal health, organ functioning, alcoholism, halitosis, drug addiction, lung inflammation, voice analysis, voice distinction, and the like. The mouthpiece 1416 is in a retracted or stored position in FIG. 14A and is in an extended or operative position in FIG. 14B.


Another multifunctional earpiece module 1500, according to embodiments of the present invention, is illustrated in FIG. 15. The illustrated earpiece module 1500 includes the embodiments described with respect to FIGS. 9 and 14A-14B, such as a pinna cover 1502, an ear support 1501, a mouthpiece 1516, an earpiece body 1505, and the like. Additionally, the earpiece module 1500 may contain an extension 1511 with sensors for monitoring jaw motion, arterial blood flow near the neck, or other physiological and environmental factors near the jaw and neck region.


The person illustrated in FIG. 15 is also wearing an earring monitor 1514 according to embodiments of the present invention. Because at least one portion of an earring may penetrate the skin, earring monitor 1514 may contain sensors and telemetric circuitry that provide access to various blood analytes through iontophoresis and electrochemical sensing that may not be easily accessible by the other portions of the earpiece module 1500. Additionally, the earring may provide a good electrical contact for ECG or skin conductivity.


Embodiments of the present invention are not limited to earpiece modules. Other types of modules may be utilized that attach to other portions of a person's body. For example, a temple module 1600 having a similar design as the earpiece module design 100 can also be employed, as illustrated in FIG. 16. A temple module 1600 has the benefit of being close to physiological areas associated with stress, intracranial pressure, brain activity, and migraines. Additionally, a temple module can monitor physiological activity associated with the onset of a stroke, such as increased or decreased blood flow and/or oxygen flow to the brain.



FIG. 19 illustrates a sensor module, according to embodiments of the present invention, integrated into a telemetric Bluetooth module. Though a Bluetooth module is used in this example, it should be understood that other telemetric modules can be used. Telemetric modules according to some embodiments of the present invention may operate in open architecture protocols, allowing multiple telemetric devices to communicate with each other. A Bluetooth module (including the sensor module) according to some embodiments of the present invention is integrated into a wearable earpiece module (i.e., module 100 described above). The sensor module illustrated in FIG. 19 contains one or more sensors, and is mounted onto a Bluetooth module. In one embodiment, the sensor module is directly soldered onto the Bluetooth module. In another embodiment, the sensor module is elevated from the Bluetooth module with spacers, and a cable or electrical wires connect between the sensor module and the Bluetooth module. The module may be elevated in embodiments where the sensors need to be exposed to the environment. For example, the sensors may need to be exposed through the frontside region 1106 of an earpiece module 1105 (FIG. 11), and the Bluetooth module may fit too deeply into the earpiece module to provide sensor access to the external environment. In some cases, contact leads or vias may connect between the sensor module and an extended sensor or an additional sensor module. This allows the extended sensor or sensor module to be flexibly mounted anywhere inside, along, outside, or about the earpiece module 100. Extended sensors can be especially useful for 4-point galvanometric monitoring of skin conductance, pulse oximetry, and volatile organic compound monitoring.



FIG. 20 illustrates the power budget of a personal health and environmental monitor earpiece module, such as earpiece module 100, incorporated into a Bluecore Bluetooth module, according to some embodiments of the present invention. As illustrated, the sensor components (i.e., a body temperature sensor, 2-axis accelerometer, barometric pressure sensor, humidity sensor, ambient temperature sensor, and ambient lighting sensor) account for less than 16 mW of the total operating power of the Bluetooth-enabled earpiece module. The Bluecore Bluetooth operating power during transmission is approximately 65 mW. Combined together, the earpiece module, with all operating components, can operate with less than 100 mW of total operating power and provide a full day of sensing between recharges of typical batteries. Sensors other than these particular sensors, can also be included with minimal increase in operating power with respect to the Bluecore Bluetooth module. Pulsed sensing or “polling” of the sensors to read out data at certain intervals can further extend battery life.


Pulse oximetry is a standard noninvasive technique of estimating blood gas levels. Pulse oximeters typically employ 2 or more optical wavelengths to estimate the ratio of oxygenated to deoxygenated blood. Similarly, various types of hemoglobin, such as methemoglobin and carboxyhemoglobin can be differentiated by measuring and comparing the optical absorption at key red and near-infrared wavelengths. FIG. 17 shows a transmittance pulse oximeter 1710 incorporated into an earpiece module (such as earpiece module 905 of FIG. 9) with the head 1707 of the earlobe clip containing an optical source 1708 and an optical detector 1709. In general, the optical wavelengths from optical source 1708 pass through arteries or veins and are selectively absorbed by various blood metabolites, typically blood gas carriers such as hemoglobin. These metabolites can change color in response to the incorporation or removal of various blood gases, such as oxygen, carbon dioxide, carbon monoxide, and other inhaled gases. The optical detector 1709 may contain optical filters to selectively detect light at key wavelengths relating to the presence or absence of the aforementioned optical absorption bands.


Though only two optical wavelengths are shown emanating from the source 1708, several additional wavelengths can be incorporated and/or replace conventional wavelengths. For example, by adding additional visible and infrared wavelengths, myoglobin, methemoglobin, carboxyhemoglobin, bilirubin, SpCO2, and blood urea nitrogen (BUN) can be estimated and/or monitored in real-time in addition to the conventional pulse oximetry SpO2 measurement.


The optical source 1708 can include light-emitting diodes (LEDs), laser diodes (LDs), or other compact optical sources. In some cases, optical energies from the optical sources can be guided with waveguides, such as fiber optics. In some cases, ambient light, such as room light or solar radiation, may be sufficient for the optical source 1708. In such case, waveguides may be used to couple ambient light towards the earlobe or other point of interest, Ambient light may be useful in that ambient light may represent a diffuse optical source that is largely independent of body position, such that motion artifacts associated with body motion may be lessened. The optical detectors 1709 can include photodiodes (PDs), avalanche photodiodes (APDs), photomultipliers, or other compact optical detectors.



FIG. 18 shows a reflective pulse oximetry setup 1800 where reflected wavelengths 1816 are measured, as opposed to measuring transmitted wavelengths. In the illustrated embodiment, an optical source-detector assembly 1811 is integrated into an earlobe clip head 1807 to generate optical wavelengths 1815 and monitor the resulting reflected optical energy 1816. The optical source-detector assembly 1811 contains one or more optical sources emitting one or more optical wavelengths, as well as one or more optical detectors detecting one or more optical wavelengths. The epidermis 1812, dermis 1813, and subcutaneous 1814 layers of skin tissue are shown in FIG. 18 for reference.


The reflective pulse oximetry setup 1817 is also suitable for measuring fluorescence from various skin or blood analytes. For example, the optical sources and/or photodetectors may be selectively filtered to measure key fluorescence bands. A fluorescence approach can be applied to, for example, the real-time monitoring of cholesterol and lipids in the skin or blood. Though the optical techniques of FIG. 17 and FIG. 18 are shown primarily over the earlobe or skin, these techniques can be integrated with the optical physiological signal extraction technique 600, described above with respect to FIG. 6, for measuring blood gas properties at or near the tympanic membrane.


Blood hydration can also be monitored optically, as water selectively absorbs optical wavelengths in the mid-IR and blue-UV ranges, whereas water can be more transparent to the blue-green wavelengths. Thus, the same optical emitter/detector configuration used in earpiece pulse oximetry (FIGS. 17 and 18) can be employed for hydration monitoring. However, mid-IR or blue optical emitters and detectors may be required. Additionally, monitoring the ratio of blue-green to other transmitted or reflected wavelengths may aid the real-time assessment of blood hydration levels. Blood hydration can also be monitored by measuring changes in capacitance, resistance, or inductance along the ear in response to varying water content in the skin tissues or blood. Similarly, hydration can be estimated by monitoring ions extracted via iontophoresis across the skin. Additionally, measuring the return velocity of reflected sound (including ultrasound) entering the head can be used to gauge hydration. These hydration sensors can be mounted anywhere within or along an earpiece. For example, with respect to the earpiece 905 of FIG. 9, hydration sensors can be mounted to a body 902 of the earpiece, the ear support 901, the earpiece backside 906, an earlobe clip, a pinna cover 1402, an earpiece fitting 1208, and the like. For monitoring hydration properties through the tympanic membrane, the earpiece tip 1212 of the earpiece fitting 1208 may be ideal for a sensor module (such as 1213). It should be noted that other hydration sensors can also be incorporated into a module.


A variety of techniques can be used for monitoring blood metabolites via an earpiece module, such as earpiece module 100. For example, glucose can be monitored via iontophoresis at the surface of the skin combined with enzyme detection. Blood urea nitrogen (BUN) can be monitored by monitoring UV fluorescence in blood (through the skin) or by monitoring visible and mid-IR light absorption using the pulse oximetry approach described above. Various ions such as sodium, potassium, magnesium, calcium, iron, copper, nickel, and other metal ions, can be monitored via selective electrodes in an earpiece module following iontophoresis through the skin. The optical physiological signal extraction approach 600 described above can be used to monitor glucose from the tympanic membrane by monitoring optical reflection and optical fluorescence from the tympanic membrane in response to IR and blue light.


Cardiopulmonary functioning can be evaluated by monitoring blood pressure, pulse, cardiac output, and blood gas levels via earpiece modules, and other monitoring apparatus in accordance with some embodiments of the present invention. Pulse rate and intensity can be monitored through pulse oximetry (described above) as well as by sensing an increase in oxygenated blood with time. Pulse rate and blood flow may also be assessed through impedance measurements via galvanometry near a blood vessel. Additionally, pulse rate and blood flow may be assessed through a fast-response thermal energy sensor, such as a pyroelectric sensor. Because moving blood may temporarily increase or decrease the localized temperature near a blood vessel, a pyroelectric sensor will generate an electrical signal that is proportional to the total blood flow in time. Blood pressure can be monitored along the earlobe, for example. According to some embodiments of the present invention, a digital blood pressure meter is integrated into an earpiece module, such as earpiece 905 of FIG. 9. A compact clip, similar to 1707 of FIG. 17, containing actuators and sonic and pressure transducers, can be placed along the earlobe, and systolic and diastolic pressure can be measured by monitoring the pressure at which the well-known Korotkoff sound is first heard (systolic), then disappears (diastolic). This technique can also be used to monitor intra-cranial pressure and other internal pressures. Blood pressure may also be measured by comparing the time between pulses at different regions of the body. For example, sensors for monitoring pulse rate and blood volume can be located in front of the ear and behind the ear or at the earlobe, and the time between the detection of each pulse from each sensor, as well as the volume of blood passed, can be processed by the signal processor 103 into an indication of blood pressure. Electrodes within or about an earpiece can also be utilized to monitor blood gases diffused through the skin, giving an indication of blood gas metabolism. For example, a compact Severinghaus electrode can be incorporated within an earpiece module for the real-time monitoring of CO2 levels in the blood, for example, through an earlobe connector, a sensor region of an earpiece fitting, or along or about an ear support. These Severinghaus-type electrodes can also be used to monitor other blood gases besides CO2, such as oxygen and nitrogen.


Organ function monitoring includes monitoring, for example, the liver, kidneys, pancreas, skin, and other vital or important organs. Liver quality can be monitored noninvasively by monitoring optical absorption and reflection at various optical wavelengths. For example, optical reflection from white LEDs or selected visible-wavelength LEDs can be used to monitor bilirubin levels in the skin and blood, for a real-time assessment of liver health.


Monitoring neurological functioning can be accomplished via electrodes placed at the ear, near the ear, or along another surface of the body. When such electrodes are placed along the forehead, this process is described as electroencephalography, and the resulting data is called an electroencephalogram (EEG). These electrodes can be either integrated into an earpiece module or connected to an earpiece module, according to some embodiments of the present invention. For example, an earlobe clip (e.g., 904, FIG. 9) can be modified to conform with EEG electrodes or other electrodes for measuring brain waves or neurological activity. For monitoring neurological functioning, a temple earpiece (e.g., 1600, FIG. 16) may also be used. Electrodes may be positioned in a temple earpiece region near the temples of a user for direct contact with the skin. In some embodiments, direct contact is not necessary, and the neurological functioning can be monitored capacitively, inductively, electromagnetically, or a combination of these approaches. In some embodiments, brain waves may couple with low frequency acoustical sensors integrated into an earpiece module.


A person's body motion and head position can be monitored by integrating a motion sensor into an earpiece module (e.g., 905, FIG. 9) Two such compact motion sensors include gyroscopes and accelerometers, typically mechanical or optical in origin. In some embodiments, an accelerometer may be composed of one or more microelectromechanical systems (MEMS) devices. In some embodiments, an accelerometer can measure acceleration or position in 2 or more axes. When the head is moved, a motion sensor detects the displaced motion from the origin. A head position monitor can be used to sense convulsions or seizures and relay this information wirelessly to a recording device. Similarly, head position monitoring may serve as a feedback mechanism for exercise and athletic training were head positioning with respect to the body is important. Additionally, the head position monitoring can be used to monitor when someone has fallen down or is not moving.


Body temperature, including core and skin temperature, can be monitored in real-time by integrating compact infrared sensors into an earpiece module, according to some embodiments of the present invention. Infrared sensors are generally composed of thermoelectric/pyroelectric materials or semiconductor devices, such as photodiodes or photoconductors. Thermistors, thermocouples, and other temperature-dependent transducers can also be incorporated for monitoring body temperature. These sensors can be very compact and thus can be integrated throughout an earpiece module. In some embodiments, these sensors may be mounted along the backside of an earpiece body, as illustrated in FIG. 9, where the earpiece connects with the ear canal. FIG. 10 shows an embodiment of a compact sensor 1009, such as a temperature sensor, incorporated into an earpiece fitting 1008 at the backside 906 of an earpiece body 902. Because the earpiece fitting 1008 is in intimate or near-intimate contact with the ear canal, body temperature can be very accurately monitored. Signal extraction technique 600, described above, may be utilized for monitoring core body temperature via the tympanic membrane.


In some embodiments of the present invention, a pedometer can be integrated into an earpiece module to measure the number of steps walked during a day. Pedometers that can be integrated into an earpiece module include, but are not limited to, mechanical pedometers (usually implementing a metallic ball or spring), microelectromechanical systems (MEMS) pedometers, inertial sensor pedometers, accelerometer-based pedometers, accelerometry, gyroscopic pedometers, and the like.


In some embodiments of the present invention, a pedometer for an earpiece module employs an acoustic sensor for monitoring the characteristic sounds of footsteps channeled along the ear canal. For example, an acoustic sensor can be integrated into an earpiece housing (e.g., 902, FIG. 9) along the backside thereof (e.g., 906, FIG. 9) and/or within an earpiece fitting thereof (e.g., 1008, FIG. 10). The sounds generated from footsteps can be detected and analyzed with a signal processor (e.g., 405, FIG. 4) using the approach described above (i.e., 500, FIG. 5) to identify footstep sounds in the midst of convoluting physiological noise. In this embodiment, digitized electrical signals from footstep sounds from outside the body are compared with digitized electrical signals from footstep sounds traveling through the body (and ear canal), and only the spectral features associated with both types of digitized signals are amplified. This provides a new signal that contains cleaner information about footsteps. Better accuracy at discriminating a true step from other sounds or motions, such as driving in a car, can be determined by analyzing more than one sensor output through the methodology 400 described in FIG. 4.


Breathing characteristics can be monitored in a manner similar to that of acoustic pedometry (described above) in the auscultatory extraction methodology 500. In some embodiments, an acoustic sensor in an earpiece module is used to sense sounds associated with breathing. Signal processing algorithms are then used to extract breathing sounds from other sounds and noise. This information is processed into a breathing monitor, capable of monitoring, for example, the intensity, volume, and speed of breathing. Another method of monitoring breathing is to employ pressure transducers into an earpiece module. Changes in pressure inside or near the ear associated with breathing can be measured directly and, through signal processing, translated into a breathing monitor. Similarly, optical reflection sensors can be used to monitor pressure in or near the ear by monitoring physical changes in the skin or tissues in response to breathing. For monitoring the physical changes of the tympanic membrane in response to breathing, and hence ascertaining breathing rate, the optical signal extraction approach 600 described above can be employed. At least one color sensor, or colormetric sensor, can be employed to monitor changes in color associated with breathing and other health factors. In the various embodiments described herein, the location of these acoustic sensors is in or near an earpiece fitting (e.g., 1008, FIG. 10) and the sensor itself is preferably positioned in a manner similar to the sensor 1009 shown in FIG. 10.


It should be noted that some embodiments of the present invention incorporate health sensors that do not employ chemical or biological reagents for monitoring various health factors. This is because such sensors have traditionally required larger instrumentation (not suitable for portability) and/or disposable samplers (not acceptable to most end users). However, sensors employing chemical or biological reagents may be incorporated into earpiece modules, according to some embodiments of the present invention. For example, the diffusion of analyte through the skin can be monitored electrically or optically by selective binding to enzymes or antibodies contained in the health sensors integrated into an earpiece module. In some cases, iontophoresis, agitation, heat, or osmosis may be required to pull ions from the skin or blood into the sensor region for monitoring health factors. In some cases, these analytes may be tagged with markers for electromagnetic, electrical, nuclear, or magnetic detection.


Caloric intake, physical activity, and metabolism can be monitored using a core temperature sensor, an accelerometer, a sound extraction methodology (e.g., 500, FIG. 5) a pulse oximeter, a hydration sensor, and the like. These sensors can be used individually or in unison to assess overall caloric metabolism and physical activity for purposes such as diet monitoring, exercise monitoring, athletic training, or the like. For example, the sound extraction methodology 500 of FIG. 5 can be used to extract sounds associated with swallowing, and this can give an indication of total food volume consumed. Additionally, a core temperature sensor, such as a thermopile, a pyroelectric sensor, a thermoelectric sensor, or a thermistor, or a tympanic membrane extraction technique (e.g., 600, FIG. 6), can be used to assess metabolism. In one case, the core temperature is compared with the outdoor temperature, and an estimate of the heat loss from the body is made, which is related to metabolism.


Environmental temperature can be monitored, for example, by thermistor, thermocouple, diode junction drop reference, or the like. Electrical temperature measurement techniques are well known to those skilled in the art, and are of suitable size and power consumption that they can be integrated into a wireless earpiece module without significant impact on the size or functionality of the wireless earpiece module.


Environmental noise can be monitored, for example, by transducer, microphone, or the like. Monitoring of environmental noise preferably includes, but is not limited to, instantaneous intensity, spectral frequency, repetition frequency, peak intensity, commonly in units of decibels, and cumulative noise level exposures, commonly in units of decibel-hours. This environmental noise may or may not include noise generated by a person wearing an earpiece module. Sound made by a person wearing an earpiece module may be filtered out, for example, using analog or digital noise cancellation techniques, by directional microphone head shaping, or the like. The environmental noise sensor may or may not be the same sensor as that used for the intended purpose of wireless communication. In some embodiments, the environmental noise sensor is a separate sensor having broader audible detection range of noise level and frequency, at the possible sacrifice of audio quality. Environmental smog includes VOC's, formaldehyde, alkenes, nitric oxide, PAH's, sulfur dioxide, carbon monoxide, olefins, aromatic compounds, xylene compounds, and the like. Monitoring of the aforementioned smog components can be performed using earpiece modules and other wearable apparatus, according to embodiments of the present invention, in a variety of methods. All smog components may be monitored. Alternatively, single smog components or combinations of smog components may be monitored. Photoionization detectors (PID's) may be used to provide continuous monitoring and instantaneous readings. Other methods of detecting smog components according to embodiments of the present invention include, but are not limited to, electrocatalytic, photocatalytic, photoelectrocatalytic, colorimetric, spectroscopic or chemical reaction methods. Examples of monitoring techniques using the aforementioned methods may include, but are not limited to, IR laser absorption spectroscopy, difference frequency generation laser spectroscopy, porous silicon optical microcavities, surface plasmon resonance, absorptive polymers, absorptive dielectrics, and colorimetric sensors. For example, absorptive polymer capacitors inductors, or other absorptive polymer-based electronics can be incorporated into an earpiece module (e.g., 100, FIG. 1) according to embodiments of the present invention. These polymers change size or electrical or optical properties in response to analyte(s) from the environment (such as those described above). The electrical signal from these absorptive polymer electronic sensors can be correlated with the type and intensity of environmental analyte. Other techniques or combinations of techniques may also be employed to monitor smog components. For example, a smog component may be monitored in addition to a reference, such as oxygen, nitrogen, hydrogen, or the like. Simultaneous monitoring of smog components with a reference analyte of known concentration allows for calibration of the estimated concentration of the smog component with respect to the reference analyte within the vicinity of an earpiece user.


In some embodiments of the present invention, environmental air particles can be monitored with a flow cell and a particle counter, particle sizer, particle identifier, or other particulate matter sensor incorporated as part of an earpiece module (e.g., 100, FIG. 1) or externally attached to an earpiece module. Non-limiting examples of particles include oil, metal shavings, dust, smoke, ash, mold, or other biological contaminates such as pollen. In some embodiments of the present invention, a sensor for monitoring particle size and concentration is an optical particle counter. A light source is used (e.g., a laser or a laser diode), to illuminate a stream of air flow. However, a directional LED beam, generated by a resonant cavity LED (RCLED), a specially lensed LED, or an intense LED point source, can also be used for particle detection. The optical detector which is off-axis from the light beam measures the amount of light scattered from a single particle by refraction and diffraction. Both the size and the number of particles can be measured at the same time. The size of the monitored particle is estimated by the intensity of the scattered light. Additionally, particles can be detected by ionization detection, as with a commercial ionization smoke detector. In this case, a low-level nuclear radiation source, such as americium-241, may be used to ionize particles in the air between two electrodes, and the total ionized charge is detected between the electrodes. As a further example, piezoelectric crystals and piezoelectric resonator devices can be used to monitor particles in that particles reaching the piezoelectric surface change the mass and hence frequency of electromechanical resonance, and this can be correlated with particle mass. If the resonators are coated with selective coatings, certain types of particles can attach preferentially to the resonator, facilitating the identification of certain types of particles in the air near a person wearing an earpiece module. In some embodiments, these resonators are solid state electrical devices, such as MEMS devices, thin film bulk acoustic resonators (FBARs), surface-acoustic wave (SAW) devices, or the like. These compact solid state components may be arrayed, each arrayed element having a different selective coating, for monitoring various types of particles.


In some embodiments of the present invention, environmental air pressure or barometric pressure can be monitored by a barometer. Non-limiting examples of barometric pressure measurement include hydrostatic columns using mercury, water, or the like, foil-based or semiconductor-based strain gauge, pressure transducers, or the like. In some embodiments of the present invention, semiconductor-based strain gauges are utilized. A strain gauge may utilize a piezoresistive material that gives an electrical response that is indicative of the amount of deflection or strain due to atmospheric pressure. Atmospheric pressure shows a diurnal cycle caused by global atmospheric tides. Environmental atmospheric pressure is of interest for prediction of weather and climate changes. Environmental pressure may also be used in conjunction with other sensing elements, such as temperature and humidity to calculate other environmental factors, such as dew point. Air pressure can also be measured by a compact MEMS device composed of a microscale diaphragm, where the diaphragm is displaced under differential pressure and this strain is monitored by the piezoelectric or piezoresistive effect.


In some embodiments of the present invention, environmental humidity, relative humidity, and dew point can be monitored by measuring capacitance, resistivity or thermal conductivity of materials exposed to the air, or by spectroscopy changes in the air itself. Resistive humidity sensors measure the change in electrical impedance of a hygroscopic medium such as a conductive polymer, salt, or treated substrate. Capacitive humidity sensors utilize incremental change in the dielectric constant of a dielectric, which is nearly directly proportional to the relative humidity of the surrounding environment. Thermal humidity sensors measure the absolute humidity by quantifying the difference between the thermal conductivity of dry air and that of air containing water vapor. Humidity data can be stored along with pressure monitor data, and a simple algorithm can be used to extrapolate the dew point. In some embodiments of the present invention, monitoring humidity is performed via spectroscopy. The absorption of light by water molecules in air is well known to those skilled in the art. The amount of absorption at known wavelengths is indicative of the humidity or relative humidity. Humidity may be monitored with a spectroscopic method that is compatible with the smog monitoring spectroscopic method described above.


When environmental factors such as the aforementioned are monitored continuously in real-time, a user's total exposure level to an environmental factor can be recorded. When a representative volume of air a user has been exposed to is monitored or estimated, the volumetric concentration of the analytes can be calculated or estimated. In order to estimate the volume of air a person wearing the earpiece has been exposed to, a pedometer or accelerometer or air flow sensor can also be integrated into an earpiece module. Pedometers and accelerometers can be integrated into an earpiece module via mechanical sensors (usually implementing a mechanical-electrical switch), MEMS devices, and/or gyroscopic technologies. The technologies required for these types of pedometers and accelerators are well known to those skilled in the art. The incorporated pedometer or accelerometer (or more than one pedometer or accelerometer) is used to gage the distance a person has traveled, for use in the estimation of the volume of air to which a person has been exposed, and the subsequent estimate of the volumetric concentration of monitored analytes.


The health and environmental sensors utilized with earpiece modules and other wearable monitoring apparatus, according to embodiments of the present invention, can operate through a user-selectable switch on an earpiece module. However, health and environmental sensors can also be run automatically and independently of the person wearing the apparatus. In other embodiments, the person may control health and environmental monitoring through a device wirelessly coupled to an earpiece module, such as a portable telecommunication device (e.g., 210, FIG. 2). For example, health and environmental sensors in or about an earpiece module can be controlled wirelessly through, for example, a cell phone, laptop, or personal digital assistant (PDA).


The earpiece module 100 may be configured such that user preferences can be “downloaded” wirelessly without requiring changes to the earpiece monitor hardware. For example, an earpiece concerned about a heart condition may wish to have the signal processor 103 focus on processing pulse signature, at the expense of ignoring other physiological or environmental parameters. The user may then use the portable telecommunication device 210 to download a specialized algorithm through the web. This may be accomplished through existing wireless infrastructure by text-messaging to a database containing the algorithm. The user will then have an earpiece module suited with analysis software specialized to the needs and desires of the user.


Health and environmental monitors, according to embodiments of the present invention, enable low-cost, real-time personal health and environmental exposure assessment monitoring of various health factors. An individual's health and environmental exposure record can be provided throughout the day, week, month, or the like. Moreover, because the health and environmental sensors can be small and compact, the overall size of an apparatus, such as an earpiece, can remain lightweight and compact.


The foregoing is illustrative of the present invention and is not to be construed as limiting thereof. Although a few exemplary embodiments of this invention have been described, those skilled in the art will readily appreciate that many modifications are possible in the exemplary embodiments without materially departing from the teachings and advantages of this invention. Accordingly, all such modifications are intended to be included within the scope of this invention as defined in the claims. The invention is defined by the following claims, with equivalents of the claims to be included therein.

Claims
  • 1. A method of monitoring a habit of a subject, the method comprising: sensing information from the subject via a plurality of sensors worn by the subject;detecting environmental information in a vicinity of the subject via at least one of the plurality of sensors worn by the subject;executing at least one algorithm via at least one signal processor in communication with the plurality of sensors, wherein the at least one algorithm is configured to: identify an occurrence of the habit via signals produced by the plurality of sensors, to log a time when the habit occurs, and to count a number of times the habit has occurred in a particular time period, wherein the habit is associated with subject physiological and/or psychological stress; andidentify the subject physiological and/or psychological stress during the occurrence of the habit;storing the detected environmental information in data storage; andaccessing the stored environmental information at a subsequent time and executing another algorithm via the at least one signal processor to determine what environmental factors in the vicinity of the subject caused the identified occurrence of the habit and the physiological and/or psychological stress associated with the habit.
  • 2. The method of claim 1, further comprising, in response to identifying the occurrence of the habit, providing audible information about the habit to the subject via at least one audible device.
  • 3. The method of claim 2, wherein the audible information comprises a warning that the habit is occurring.
  • 4. The method of claim 1, wherein the plurality of sensors are supported by an apparatus configured to be worn at an ear or head of the subject.
  • 5. The method of claim 1, wherein at least one of the plurality of sensors comprises an acoustical sensor, and wherein sensing information from the subject comprises detecting sounds from the subject via the acoustical sensor.
  • 6. The method of claim 4, wherein the apparatus comprises the at least one signal processor.
  • 7. The method of claim 2, wherein the plurality of sensors, the at least one signal processor, and the at least one audible device are supported by an apparatus configured to worn at an ear or head of the subject.
  • 8. The method of claim 1, wherein the environmental factors include one or more of the following: noise, external sounds, lighting condition, light intensity and color, ozone level, ambient temperature.
  • 9. The method of claim 1, wherein the another algorithm is configured to determine whether noise, external sounds, lighting condition, light intensity and color, ozone level, or ambient temperature caused the identified subject physiological and/or psychological stress.
  • 10. The method of claim 1, wherein the at least one algorithm is configured to identify the subject physiological and/or psychological stress from one or more of the following obtained from the plurality of sensors: subject muscle tension, subject breathing rate, subject pulse rate, subject neurological activity, subject brain wave intensity and activity, subject core body temperature, subject blood pressure, subject galvanometric response.
  • 11. The method of claim 1, wherein the at least one algorithm comprises two or more algorithms.
  • 12. A method of monitoring a habit of a subject, the method comprising: sensing information from the subject via a plurality of sensors worn by the subject;detecting environmental information in a vicinity of the subject via at least one of the plurality of sensors worn by the subject;executing at least one algorithm via at least one signal processor in communication with the plurality of sensors, wherein the at least one algorithm is configured to: identify an occurrence of the habit via signals produced by the plurality of sensors, to log a time when the habit occurs, and to count a number of times the habit has occurred in a particular time period, wherein the habit is associated with subject physiological and/or psychological stress; andidentify the subject physiological and/or psychological stress during the occurrence of the habit from one or more of the following sensed by the plurality of sensors: subject muscle tension, subject breathing rate, subject pulse rate, subject neurological activity, subject brain wave intensity and activity, subject core body temperature, subject blood pressure, subject galvanometric response;storing the detected environmental information in data storage; andaccessing the stored environmental information at a subsequent time and executing another algorithm via the at least one signal processor to determine whether one or more of the following environmental factors in the vicinity of the subject caused the identified occurrence of the habit and the physiological and/or psychological stress associated with the habit: noise, external sounds, lighting condition, light intensity and color, ozone level, or ambient temperature.
  • 13. The method of claim 12, further comprising, in response to identifying the occurrence of the habit, providing audible information about the habit to the subject via at least one audible device, wherein the audible information comprises a warning that the habit is occurring.
  • 14. The method of claim 12, wherein the plurality of sensors are supported by an apparatus configured to be worn at an ear or head of the subject.
  • 15. The method of claim 12, wherein at least one of the plurality of sensors comprises an acoustical sensor, and wherein sensing information from the subject comprises detecting sounds from the subject via the acoustical sensor.
  • 16. The method of claim 14, wherein the apparatus comprises the at least one signal processor.
  • 17. The method of claim 13, wherein the plurality of sensors, the at least one signal processor, and the at least one audible device are supported by an apparatus configured to worn at an ear or head of the subject.
  • 18. The method of claim 12, wherein the at least one algorithm comprises two or more algorithms.
RELATED APPLICATION

This application is a divisional application of pending U.S. patent application Ser. No. 14/328,107, filed Jul. 10, 2014, which is a continuation application of pending U.S. patent application Ser. No. 14/063,669, filed Oct. 25, 2013, which is a continuation application of U.S. patent application Ser. No. 11/811,844, filed Jun. 12, 2007, now U.S. Pat. No. 8,652,040, and which claims the benefit of and priority to U.S. Provisional Patent Application No. 60/905,761, filed Mar. 8, 2007, U.S. Provisional Patent Application No. 60/876,128, filed Dec. 21, 2006, and U.S. Provisional Patent Application No. 60/875,606, filed Dec. 19, 2006, the disclosures of which are incorporated herein by reference as if set forth in their entireties.

US Referenced Citations (580)
Number Name Date Kind
3595219 Friedlander et al. Jul 1971 A
3922488 Gabr Nov 1975 A
4005701 Aisenberg et al. Feb 1977 A
4025734 Aloupis May 1977 A
4240882 Ang et al. Dec 1980 A
4312358 Barney Jan 1982 A
4331154 Broadwater et al. May 1982 A
4438772 Slavin Mar 1984 A
4459645 Glatter Jul 1984 A
4491760 Linvill Jan 1985 A
4521499 Switzer Jun 1985 A
4541905 Kuwana et al. Sep 1985 A
4592807 Switzer Jun 1986 A
4598417 Deno Jul 1986 A
4655225 Dahne et al. Apr 1987 A
4736431 Allie et al. Apr 1988 A
4783815 Büttner Nov 1988 A
4830014 Goodman et al. May 1989 A
4850962 Schaefer Jul 1989 A
4878501 Shue Nov 1989 A
4882492 Schlager Nov 1989 A
4896676 Sasaki Jan 1990 A
4928704 Hardt May 1990 A
4952890 Swanson Aug 1990 A
4952928 Carroll et al. Aug 1990 A
4957109 Groeger et al. Sep 1990 A
4985925 Langberg Jan 1991 A
5002060 Nedivi Mar 1991 A
5022970 Cook et al. Jun 1991 A
5025791 Niwa Jun 1991 A
5046103 Warnaka et al. Sep 1991 A
5079421 Knudson et al. Jan 1992 A
5080098 Willett et al. Jan 1992 A
5086229 Rosenthal et al. Feb 1992 A
5091954 Sasaki et al. Feb 1992 A
5115133 Knudson May 1992 A
5119819 Thomas et al. Jun 1992 A
5131047 Hashimoto et al. Jul 1992 A
5138663 Moseley Aug 1992 A
5139025 Lewis et al. Aug 1992 A
5143078 Mather et al. Sep 1992 A
5179951 Knudson Jan 1993 A
5182774 Bourk Jan 1993 A
5226417 Swedlow et al. Jul 1993 A
5237994 Goldberger Aug 1993 A
5259033 Goodings et al. Nov 1993 A
5299570 Hatschek Apr 1994 A
5309922 Schechter et al. May 1994 A
5348002 Caro Sep 1994 A
5377100 Pope et al. Dec 1994 A
5402496 Soli et al. Mar 1995 A
5444786 Raviv Aug 1995 A
5448082 Kim Sep 1995 A
5467775 Callahan et al. Nov 1995 A
5469855 Pompei et al. Nov 1995 A
5471009 Oba et al. Nov 1995 A
5481615 Eatwell et al. Jan 1996 A
5482036 Diab et al. Jan 1996 A
5492129 Greenberger Feb 1996 A
5494043 O'Sullivan et al. Feb 1996 A
5499301 Sudo et al. Mar 1996 A
5539831 Harley Jul 1996 A
5572990 Berlin Nov 1996 A
5581648 Sahagen Dec 1996 A
5596987 Chance Jan 1997 A
5652570 Lepkofker Jul 1997 A
5662117 Bittman Sep 1997 A
5671301 Kupershmidt Sep 1997 A
5673692 Schulze et al. Oct 1997 A
5697374 Odagiri et al. Dec 1997 A
5704365 Albrecht et al. Jan 1998 A
5711308 Singer Jan 1998 A
5721783 Anderson Feb 1998 A
5722418 Bro Mar 1998 A
5725480 Oosta et al. Mar 1998 A
5743260 Chung et al. Apr 1998 A
5779631 Chance Jul 1998 A
5797841 Delonzor et al. Aug 1998 A
5807114 Hodges et al. Sep 1998 A
5807267 Bryars et al. Sep 1998 A
5817008 Rafert et al. Oct 1998 A
5846190 Woehrle Dec 1998 A
5853005 Scanlon Dec 1998 A
5873836 Kahn et al. Feb 1999 A
5881159 Aceti Mar 1999 A
5904654 Wohltmann et al. May 1999 A
5938593 Quellette Aug 1999 A
5954644 Dettling et al. Sep 1999 A
5964701 Asada et al. Oct 1999 A
5971931 Raff Oct 1999 A
5974338 Asano et al. Oct 1999 A
5995858 Kinast Nov 1999 A
6004274 Aceti et al. Dec 1999 A
6013007 Root et al. Jan 2000 A
6022748 Charych et al. Feb 2000 A
6023541 Merchant et al. Feb 2000 A
6030342 Amano et al. Feb 2000 A
6045511 Ott et al. Apr 2000 A
6052336 Lowrey, III Apr 2000 A
6067006 O'Brien May 2000 A
6070093 Oosta et al. May 2000 A
6078829 Uchida et al. Jun 2000 A
6080110 Thorgersen Jun 2000 A
6081742 Amano et al. Jun 2000 A
6144867 Walker et al. Nov 2000 A
6155983 Kosuda et al. Dec 2000 A
6168567 Pickering et al. Jan 2001 B1
6186145 Brown Feb 2001 B1
6198394 Jacobsen et al. Mar 2001 B1
6198951 Kosuda et al. Mar 2001 B1
6205354 Gellermann et al. Mar 2001 B1
6231519 Blants et al. May 2001 B1
6253871 Aceti Jul 2001 B1
6267721 Welles Jul 2001 B1
6277079 Avicola et al. Aug 2001 B1
6283915 Nolan et al. Sep 2001 B1
6285816 Anderson et al. Sep 2001 B1
6289230 Chaiken et al. Sep 2001 B1
6298314 Blackadar et al. Oct 2001 B1
6304797 Shusterman Oct 2001 B1
6332868 Sato et al. Dec 2001 B1
6340350 Simms Jan 2002 B1
6358216 Kraus et al. Mar 2002 B1
6361660 Goldstein Mar 2002 B1
6371925 Imai et al. Apr 2002 B1
6373942 Braund Apr 2002 B1
6374129 Chin et al. Apr 2002 B1
6385176 Iyengar et al. May 2002 B1
6409675 Turcott Jun 2002 B1
6415167 Blank et al. Jul 2002 B1
6443890 Schulze et al. Sep 2002 B1
6444474 Thomas et al. Sep 2002 B1
6445799 Taenzer et al. Sep 2002 B1
6454718 Clift Sep 2002 B1
6458080 Brown et al. Oct 2002 B1
6470893 Boesen Oct 2002 B1
6491644 Vujanic et al. Dec 2002 B1
6491647 Bridger et al. Dec 2002 B1
6513532 Mault et al. Feb 2003 B2
6514278 Hibst et al. Feb 2003 B1
6527711 Stivoric et al. Mar 2003 B1
6527712 Brown et al. Mar 2003 B1
6527729 Turcott Mar 2003 B1
6529754 Kondo Mar 2003 B2
6534012 Hazen et al. Mar 2003 B1
6544199 Morris Apr 2003 B1
6556852 Schulze et al. Apr 2003 B1
6569094 Suzuki et al. May 2003 B2
6571117 Marbach May 2003 B1
6605038 Teller et al. Aug 2003 B1
6608562 Kimura et al. Aug 2003 B1
6616606 Petersen et al. Sep 2003 B1
6616613 Goodman Sep 2003 B1
6631196 Taenzer et al. Oct 2003 B1
6647378 Kindo Nov 2003 B2
6656116 Kim et al. Dec 2003 B2
6694180 Boesen Feb 2004 B1
6702752 Dekker Mar 2004 B2
6725072 Steuer et al. Apr 2004 B2
6738485 Boesen May 2004 B1
6745061 Hicks et al. Jun 2004 B1
6748254 O'Neil et al. Jun 2004 B2
6760610 Tschupp et al. Jul 2004 B2
6783501 Takahashi et al. Aug 2004 B2
6808473 Hisano et al. Oct 2004 B2
6859658 Krug Feb 2005 B1
6893396 Schulze et al. May 2005 B2
6941239 Unuma et al. Sep 2005 B2
6953435 Kondo et al. Oct 2005 B2
6954644 Johansson et al. Oct 2005 B2
6996427 Ali et al. Feb 2006 B2
6997879 Turcott Feb 2006 B1
7011814 Suddarth Mar 2006 B2
7018338 Vetter et al. Mar 2006 B2
7024369 Brown et al. Apr 2006 B1
7030359 Römhild Apr 2006 B2
7034694 Yamaguchi et al. Apr 2006 B2
7041062 Friedrichs et al. May 2006 B2
7043287 Khalil et al. May 2006 B1
7054674 Cane et al. May 2006 B2
7088234 Naito et al. Aug 2006 B2
7088828 Bradford et al. Aug 2006 B1
7107088 Aceti Sep 2006 B2
7113815 O'Neil et al. Sep 2006 B2
7117032 Childre et al. Oct 2006 B2
7163512 Childre et al. Jan 2007 B1
7175601 Verjus et al. Feb 2007 B2
7190986 Hannula et al. Mar 2007 B1
7209775 Bae et al. Apr 2007 B2
7217224 Thomas May 2007 B2
7252639 Kimura et al. Aug 2007 B2
7263396 Chen et al. Aug 2007 B2
7289837 Mannheimer et al. Oct 2007 B2
7324668 Rubinstenn Jan 2008 B2
7336982 Yoo et al. Feb 2008 B2
7341559 Schultz et al. Mar 2008 B2
7376451 Mahony et al. May 2008 B2
7378954 Wendt May 2008 B2
7470234 Elhag et al. Dec 2008 B1
7483730 Diab et al. Jan 2009 B2
7486988 Goodall et al. Feb 2009 B2
7507207 Sakai et al. Mar 2009 B2
7519327 White Apr 2009 B2
7526327 Blondeau et al. Apr 2009 B2
7558625 Levin et al. Jul 2009 B2
7583994 Scholz Sep 2009 B2
7620450 Kim et al. Nov 2009 B2
7625285 Breving Dec 2009 B2
7652569 Kiff et al. Jan 2010 B2
7689437 Teller et al. Mar 2010 B1
7695440 Kondo et al. Apr 2010 B2
7725147 Li et al. May 2010 B2
7756559 Abreu Jul 2010 B2
7843325 Otto Nov 2010 B2
7881733 Staton et al. Feb 2011 B2
7894869 Hoarau Feb 2011 B2
7914468 Shalon et al. Mar 2011 B2
7991448 Edgar et al. Aug 2011 B2
7998079 Nagai et al. Aug 2011 B2
8024974 Bharti et al. Sep 2011 B2
8050728 Al-Ali et al. Nov 2011 B2
8055319 Oh et al. Nov 2011 B2
8055330 Egozi Nov 2011 B2
8059924 Letant et al. Nov 2011 B1
8130105 Al-Ali et al. Mar 2012 B2
8137270 Keenan et al. Mar 2012 B2
8172459 Abreu May 2012 B2
8175670 Baker, Jr. et al. May 2012 B2
8204730 Liu et al. Jun 2012 B2
8233955 Ai-Aii et al. Jul 2012 B2
8251903 LeBoeuf et al. Aug 2012 B2
8255027 Al-Ali et al. Aug 2012 B2
8255029 Addison et al. Aug 2012 B2
8303512 Kosuda et al. Nov 2012 B2
8328420 Abreu Dec 2012 B2
8385560 Solbeck et al. Feb 2013 B2
8416959 Lott et al. Apr 2013 B2
8491492 Shinar et al. Jul 2013 B2
8504679 Spire et al. Aug 2013 B2
8506524 Graskov et al. Aug 2013 B2
8512242 LeBoeuf et al. Aug 2013 B2
8679008 Hughes et al. Mar 2014 B2
8730048 Shen et al. May 2014 B2
9005129 Venkatraman et al. Apr 2015 B2
20010000526 Gopinathan et al. Apr 2001 A1
20010015123 Nishitani et al. Aug 2001 A1
20010027384 Schulze et al. Oct 2001 A1
20010039372 Yasushi et al. Nov 2001 A1
20010040591 Abbott et al. Nov 2001 A1
20010044588 Mault Nov 2001 A1
20010049471 Suzuki et al. Dec 2001 A1
20010051766 Gazdzinski Dec 2001 A1
20020019586 Teller Feb 2002 A1
20020021800 Bodley Feb 2002 A1
20020035340 Fraden et al. Mar 2002 A1
20020099356 Unger et al. Jul 2002 A1
20020107649 Takiguchi et al. Aug 2002 A1
20020115937 Song Aug 2002 A1
20020143242 Nemirovski Oct 2002 A1
20020156386 Dardik et al. Oct 2002 A1
20020156654 Roe et al. Oct 2002 A1
20020165466 Givens Nov 2002 A1
20020169485 Pless et al. Nov 2002 A1
20020180605 Ozguz et al. Dec 2002 A1
20020186137 Skardon Dec 2002 A1
20020188210 Aizawa Dec 2002 A1
20020194002 Petrushin Dec 2002 A1
20030002685 Werblud Jan 2003 A1
20030002705 Boesen Jan 2003 A1
20030007631 Bolognesi et al. Jan 2003 A1
20030036685 Goodman Feb 2003 A1
20030045785 Diab et al. Mar 2003 A1
20030050563 Suribhotla et al. Mar 2003 A1
20030060693 Monfre et al. Mar 2003 A1
20030064712 Gaston et al. Apr 2003 A1
20030065257 Mault et al. Apr 2003 A1
20030065269 Vetter et al. Apr 2003 A1
20030083583 Kovtun et al. May 2003 A1
20030109030 Uchida et al. Jun 2003 A1
20030109791 Kondo et al. Jun 2003 A1
20030118197 Nagayasu et al. Jun 2003 A1
20030147369 Singh et al. Aug 2003 A1
20030147544 Lichtblau Aug 2003 A1
20030149526 Zhou et al. Aug 2003 A1
20030151524 Clark Aug 2003 A1
20030163710 Ortiz et al. Aug 2003 A1
20030181795 Suzuki et al. Sep 2003 A1
20030181798 Al-Ali Sep 2003 A1
20030195040 Breving Oct 2003 A1
20030208113 Mault et al. Nov 2003 A1
20030212336 Lee et al. Nov 2003 A1
20030220584 Honeyager et al. Nov 2003 A1
20030222268 Yocom et al. Dec 2003 A1
20030233051 Verjus et al. Dec 2003 A1
20030234726 Chen et al. Dec 2003 A1
20040004547 Appelt et al. Jan 2004 A1
20040022700 Kim et al. Feb 2004 A1
20040030581 Leven Feb 2004 A1
20040032957 Mansy et al. Feb 2004 A1
20040034289 Teller et al. Feb 2004 A1
20040034293 Kimball Feb 2004 A1
20040039254 Stivoric et al. Feb 2004 A1
20040054291 Schulz et al. Mar 2004 A1
20040075677 Loyall et al. Apr 2004 A1
20040077934 Massad Apr 2004 A1
20040082842 Lumba et al. Apr 2004 A1
20040092846 Watrous May 2004 A1
20040103146 Park May 2004 A1
20040117204 Mazar et al. Jun 2004 A1
20040120844 Tribelsky et al. Jun 2004 A1
20040122294 Hatlestad et al. Jun 2004 A1
20040122702 Sabol et al. Jun 2004 A1
20040133123 Leonhardt et al. Jul 2004 A1
20040135571 Uutela et al. Jul 2004 A1
20040138578 Pineda et al. Jul 2004 A1
20040183675 Harris Sep 2004 A1
20040186387 Kosuda et al. Sep 2004 A1
20040186390 Ross et al. Sep 2004 A1
20040198463 Knoedgen Oct 2004 A1
20040203897 Rogers Oct 2004 A1
20040212505 Dewing et al. Oct 2004 A1
20040215958 Ellis et al. Oct 2004 A1
20040219056 Tribelsky et al. Nov 2004 A1
20040220483 Yeo et al. Nov 2004 A1
20040220488 Vyshedskiy et al. Nov 2004 A1
20040225207 Bae et al. Nov 2004 A1
20040228494 Smith Nov 2004 A1
20040240516 Harr Dec 2004 A1
20040242976 Abreu Dec 2004 A1
20040254501 Mault Dec 2004 A1
20050004458 Kanayama et al. Jan 2005 A1
20050007582 Villers et al. Jan 2005 A1
20050021519 Ghouri Jan 2005 A1
20050027216 Guillemaud et al. Feb 2005 A1
20050030540 Thornton Feb 2005 A1
20050033200 Soehren et al. Feb 2005 A1
20050036212 Saito Feb 2005 A1
20050038349 Choi et al. Feb 2005 A1
20050043600 Diab et al. Feb 2005 A1
20050043630 Honeyager et al. Feb 2005 A1
20050058456 Yoo Mar 2005 A1
20050059870 Aceti Mar 2005 A1
20050084666 Pong et al. Apr 2005 A1
20050101845 Nihtila May 2005 A1
20050101872 Sattler et al. May 2005 A1
20050113167 Buchner et al. May 2005 A1
20050113656 Chance May 2005 A1
20050113703 Farringdon et al. May 2005 A1
20050116820 Goldreich Jun 2005 A1
20050119833 Nanikashvili Jun 2005 A1
20050134452 Smith Jun 2005 A1
20050148883 Boesen Jul 2005 A1
20050154264 Lecompte et al. Jul 2005 A1
20050163302 Mock et al. Jul 2005 A1
20050177029 Shen Aug 2005 A1
20050177034 Beaumont Aug 2005 A1
20050187448 Petersen et al. Aug 2005 A1
20050187453 Petersen et al. Aug 2005 A1
20050190072 Brown et al. Sep 2005 A1
20050192514 Kearby et al. Sep 2005 A1
20050192515 Givens et al. Sep 2005 A1
20050192516 Takiguchi et al. Sep 2005 A1
20050192557 Brauker et al. Sep 2005 A1
20050196009 Boesen Sep 2005 A1
20050203349 Nanikashvili Sep 2005 A1
20050203357 Debreczeny et al. Sep 2005 A1
20050209516 Fraden Sep 2005 A1
20050212405 Negley Sep 2005 A1
20050222487 Miller et al. Oct 2005 A1
20050222903 Buchheit et al. Oct 2005 A1
20050226446 Luo et al. Oct 2005 A1
20050228244 Banet Oct 2005 A1
20050228299 Banet Oct 2005 A1
20050228463 Mac et al. Oct 2005 A1
20050240087 Keenan et al. Oct 2005 A1
20050245839 Stivoric et al. Nov 2005 A1
20050258816 Zen et al. Nov 2005 A1
20050258950 Sahashi et al. Nov 2005 A1
20050259811 Kimm et al. Nov 2005 A1
20060004680 Robarts et al. Jan 2006 A1
20060009685 Finarov et al. Jan 2006 A1
20060012567 Sicklinger Jan 2006 A1
20060047215 Newman et al. Mar 2006 A1
20060061468 Ruha Mar 2006 A1
20060063993 Yu et al. Mar 2006 A1
20060064037 Shalon Mar 2006 A1
20060073807 Baker et al. Apr 2006 A1
20060075257 Martis et al. Apr 2006 A1
20060084878 Banet et al. Apr 2006 A1
20060084879 Nazarian et al. Apr 2006 A1
20060122520 Banet et al. Jun 2006 A1
20060123885 Yates et al. Jun 2006 A1
20060140425 Berg et al. Jun 2006 A1
20060142665 Garay et al. Jun 2006 A1
20060202816 Crump et al. Sep 2006 A1
20060205083 Zhao Sep 2006 A1
20060206014 Ariav Sep 2006 A1
20060210058 Kock et al. Sep 2006 A1
20060211922 Al-Ali et al. Sep 2006 A1
20060211924 Dalke et al. Sep 2006 A1
20060212316 Jackson Sep 2006 A1
20060217598 Miyajima et al. Sep 2006 A1
20060217988 Sukkar et al. Sep 2006 A1
20060224059 Swedlow et al. Oct 2006 A1
20060226991 Rivas Oct 2006 A1
20060240558 Zhao Oct 2006 A1
20060246342 MacPhee Nov 2006 A1
20060251277 Cho Nov 2006 A1
20060251334 Oba et al. Nov 2006 A1
20060252999 Devaul et al. Nov 2006 A1
20060264730 Stivoric et al. Nov 2006 A1
20060292533 Selod Dec 2006 A1
20060293839 Stankieiwcz et al. Dec 2006 A1
20060293921 McCarthy et al. Dec 2006 A1
20070004449 Sham Jan 2007 A1
20070004969 Kong et al. Jan 2007 A1
20070015992 Filkins et al. Jan 2007 A1
20070021206 Sunnen Jan 2007 A1
20070027367 Oliver et al. Feb 2007 A1
20070027399 Chou Feb 2007 A1
20070036383 Romero Feb 2007 A1
20070043304 Katayama Feb 2007 A1
20070050215 Kil et al. Mar 2007 A1
20070060800 Drinan et al. Mar 2007 A1
20070060819 Altschuler et al. Mar 2007 A1
20070063850 Devaul et al. Mar 2007 A1
20070082789 Nissila et al. Apr 2007 A1
20070083092 Rippo et al. Apr 2007 A1
20070083095 Rippo et al. Apr 2007 A1
20070088221 Stahmann Apr 2007 A1
20070093702 Yu et al. Apr 2007 A1
20070106167 Kinast May 2007 A1
20070112273 Rogers May 2007 A1
20070112277 Fischer et al. May 2007 A1
20070112598 Heckerman et al. May 2007 A1
20070116314 Grilliot et al. May 2007 A1
20070118054 Oliver et al. May 2007 A1
20070123763 Al-Ali et al. May 2007 A1
20070135717 Uenishi et al. Jun 2007 A1
20070135866 Baker et al. Jun 2007 A1
20070159926 Prstojevich Jul 2007 A1
20070165872 Bridger et al. Jul 2007 A1
20070167850 Russell et al. Jul 2007 A1
20070179739 Donofrio et al. Aug 2007 A1
20070191718 Nakamura Aug 2007 A1
20070197878 Shklarski Aug 2007 A1
20070197881 Wolf et al. Aug 2007 A1
20070213020 Novac Sep 2007 A1
20070230714 Armstrong Oct 2007 A1
20070233403 Alwan et al. Oct 2007 A1
20070265097 Havukainen Nov 2007 A1
20070265508 Sheikhzadeh-Nadjar et al. Nov 2007 A1
20070270667 Coppi et al. Nov 2007 A1
20070270671 Gal Nov 2007 A1
20070273504 Tran Nov 2007 A1
20070276270 Tran Nov 2007 A1
20070293781 Sims et al. Dec 2007 A1
20070299330 Couronne et al. Dec 2007 A1
20080001735 Tran Jan 2008 A1
20080004536 Baxi et al. Jan 2008 A1
20080004904 Tran Jan 2008 A1
20080015424 Bernreuter Jan 2008 A1
20080039731 McCombie et al. Feb 2008 A1
20080076972 Dorogusker et al. Mar 2008 A1
20080081963 Naghavi et al. Apr 2008 A1
20080081972 Debreczeny Apr 2008 A1
20080086533 Neuhauser et al. Apr 2008 A1
20080096726 Riley et al. Apr 2008 A1
20080106404 Joslin et al. May 2008 A1
20080114220 Banet et al. May 2008 A1
20080132798 Hong et al. Jun 2008 A1
20080133699 Craw et al. Jun 2008 A1
20080141301 Azzaro et al. Jun 2008 A1
20080146890 LeBoeuf et al. Jun 2008 A1
20080146892 LeBoeuf et al. Jun 2008 A1
20080154098 Morris et al. Jun 2008 A1
20080154105 Lemay Jun 2008 A1
20080165017 Schwartz Jul 2008 A1
20080170600 Sattler et al. Jul 2008 A1
20080171945 Dotter Jul 2008 A1
20080177162 Bae et al. Jul 2008 A1
20080187163 Goldstein et al. Aug 2008 A1
20080200774 Luo Aug 2008 A1
20080203144 Kim Aug 2008 A1
20080221461 Zhou et al. Sep 2008 A1
20080249594 Dietrich Oct 2008 A1
20080287752 Stroetz et al. Nov 2008 A1
20080298624 Jeong et al. Dec 2008 A1
20090005662 Petersen et al. Jan 2009 A1
20090006457 Stivoric et al. Jan 2009 A1
20090010461 Klinghult et al. Jan 2009 A1
20090010556 Uchibayashi et al. Jan 2009 A1
20090024004 Yang Jan 2009 A1
20090030350 Yang et al. Jan 2009 A1
20090034748 Sibbald Feb 2009 A1
20090048498 Riskey Feb 2009 A1
20090054751 Babashan et al. Feb 2009 A1
20090054752 Jonnalagadda et al. Feb 2009 A1
20090069645 Nielsen et al. Mar 2009 A1
20090082994 Schuler et al. Mar 2009 A1
20090088611 Buschmann Apr 2009 A1
20090093687 Telfort et al. Apr 2009 A1
20090097681 Puria et al. Apr 2009 A1
20090105548 Bart Apr 2009 A1
20090105556 Fricke et al. Apr 2009 A1
20090112071 LeBoeuf et al. Apr 2009 A1
20090131761 Moroney, III et al. May 2009 A1
20090131764 Lee et al. May 2009 A1
20090171221 Liao et al. Jul 2009 A1
20090175456 Johnson Jul 2009 A1
20090177097 Ma et al. Jul 2009 A1
20090214060 Chuang et al. Aug 2009 A1
20090221888 Wijesiriwardana Sep 2009 A1
20090227853 Wijesiriwardana Sep 2009 A1
20090240125 Such et al. Sep 2009 A1
20090253992 Van Der Loo Oct 2009 A1
20090253996 Lee et al. Oct 2009 A1
20090264711 Schuler et al. Oct 2009 A1
20090268911 Singh Oct 2009 A1
20090270698 Shioi et al. Oct 2009 A1
20090281435 Ahmed et al. Nov 2009 A1
20090287067 Dorogusker et al. Nov 2009 A1
20090299215 Zhang Dec 2009 A1
20100004517 Bryenton et al. Jan 2010 A1
20100004860 Chernoguz et al. Jan 2010 A1
20100022861 Cinbis et al. Jan 2010 A1
20100045663 Chen et al. Feb 2010 A1
20100100013 Hu et al. Apr 2010 A1
20100113948 Yang et al. May 2010 A1
20100168531 Shaltis et al. Jul 2010 A1
20100172510 Juvonen Jul 2010 A1
20100172522 Mooring et al. Jul 2010 A1
20100179389 Moroney et al. Jul 2010 A1
20100185105 Baldinger Jul 2010 A1
20100217102 LeBoeuf et al. Aug 2010 A1
20100217103 Abdul-Hafiz et al. Aug 2010 A1
20100222655 Starr et al. Sep 2010 A1
20100228315 Nielsen Sep 2010 A1
20100234714 Mercier et al. Sep 2010 A1
20100268056 Picard et al. Oct 2010 A1
20100274100 Behar et al. Oct 2010 A1
20100274109 Hu et al. Oct 2010 A1
20100292589 Goodman Nov 2010 A1
20100298653 McCombie et al. Nov 2010 A1
20110028810 Van Slyke et al. Feb 2011 A1
20110028813 Watson et al. Feb 2011 A1
20110081037 Oh et al. Apr 2011 A1
20110105869 Wilson et al. May 2011 A1
20110112382 Li et al. May 2011 A1
20110130638 Raridan, Jr. Jun 2011 A1
20110142371 King et al. Jun 2011 A1
20110288379 Wu Nov 2011 A1
20120030547 Raptis et al. Feb 2012 A1
20120039493 Rucker et al. Feb 2012 A1
20120095303 He Apr 2012 A1
20120156933 Kreger et al. Jun 2012 A1
20120179011 Moon et al. Jul 2012 A1
20120197093 LeBoeuf et al. Aug 2012 A1
20120277548 Burton Nov 2012 A1
20130053661 Alberth et al. Feb 2013 A1
20130063550 Ritchey et al. Mar 2013 A1
20130072765 Kahn et al. Mar 2013 A1
20130131519 LeBoeuf et al. May 2013 A1
20130218588 Kehr et al. Aug 2013 A1
20130245387 Patel Sep 2013 A1
20130336495 Burgett et al. Dec 2013 A1
20140051940 Messerschmidt Feb 2014 A1
20140052567 Bhardwaj et al. Feb 2014 A1
20140073486 Ahmed et al. Mar 2014 A1
20140100432 Golda et al. Apr 2014 A1
20140127996 Park et al. May 2014 A1
20140203797 Stivoric et al. Jul 2014 A1
20140219467 Kurtz Aug 2014 A1
20140221777 Betts Aug 2014 A1
20140236531 Carter Aug 2014 A1
20140275852 Hong et al. Sep 2014 A1
20140323880 Ahmed et al. Oct 2014 A1
20140378844 Fei Dec 2014 A1
20160287108 Wei et al. Oct 2016 A1
20170034615 Mankodi et al. Feb 2017 A1
Foreign Referenced Citations (44)
Number Date Country
101212927 Jul 2008 CN
201438747 Apr 2010 CN
3910749 Oct 1990 DE
1 297 784 Apr 2003 EP
1 480 278 Nov 2004 EP
2 077 091 Jul 2009 EP
2 182 839 Oct 2011 EP
2 408 209 May 2005 GB
2 411 719 Sep 2005 GB
7-241279 Sep 1995 JP
9-253062 Sep 1997 JP
9-299342 Nov 1997 JP
2000-116611 Apr 2000 JP
2001-025462 Jan 2001 JP
20030159221 Jun 2003 JP
2004-513750 May 2004 JP
2004-283523 Oct 2004 JP
2005-040261 Feb 2005 JP
2005-270544 Oct 2005 JP
2007-044203 Feb 2007 JP
2007-185348 Jul 2007 JP
2008-136556 Jun 2008 JP
2008-279061 Nov 2008 JP
2009-153664 Jul 2009 JP
2010-526646 Aug 2010 JP
2014-068733 Apr 2014 JP
20-0204510 Nov 2000 KR
WO 0024064 Apr 2000 WO
WO 2000047108 Aug 2000 WO
WO 0108552 Feb 2001 WO
WO 0217782 Mar 2002 WO
WO 2005010568 Feb 2005 WO
WO 2005015163 Feb 2005 WO
WO 2005020121 Mar 2005 WO
WO 2005036212 Apr 2005 WO
WO 2005074550 Aug 2005 WO
WO 2005110238 Nov 2005 WO
WO 2006009830 Jan 2006 WO
WO 2006067690 Jun 2006 WO
WO 2007012931 Feb 2007 WO
WO 2007053146 May 2007 WO
WO 2008141306 Nov 2008 WO
WO 2011127063 Oct 2011 WO
WO 2013038296 Mar 2013 WO
Non-Patent Literature Citations (104)
Entry
Anpo et al. “Photocatalytic Reduction of Co2 With H2O on Titanium Oxides Anchored within Micropores of Zeolites: Effects of the Structure of the Active Sites and the Addition of Pt” J. Phys. Chem. B, 101:2632-2636 (1997).
Bârsan et al. “Understanding the fundamental principles of metal oxide based gas sensors; the example of CO sensing with SnO2 sensors in the presence of humidity” Journal of Physics: Condensed Matter 15:R813-R839 (2003).
Bott “Electrochemistry of Semiconductors” Current Separations 17(3):87-91 (1998).
European Search Report corresponding to European Application No. 07862660.3 dated Apr. 25, 2012; 7 pages.
Fitrainer “The Only Trainer You Need”; http://itami.com; Downloaded Feb. 26, 2010; © 2008 FiTriainer™; 2 pages.
International Search Report and Written Opinion of the International Searching Authority, corresponding to PCT/US2007/025114, dated May 13, 2008.
Martins et al. “Zinc oxide as an ozone sensor” Journal of Applied Physics 96(3):1398-1408 (2004).
Saladin et al. “Photosynthesis of CH4 at a TiO2 Surface from Gaseous H2O and CO2” J. Chem. Soc., Chem. Commun. 533-534 (1995).
Skubal et al. “Detection and identification of gaseous organics using a TiO2 sensor” Journal of Photochemistry and Photobiology A: Chemistry 148:103-108 (2002).
Skubal et al. “Monitoring the Electrical Response of Photoinduced Organic Oxideation on TiO2 Surfaces” Manuscript submitted Oct. 2000 to SPIE Intl. Symposium on Environment & Industrial Sensing, Boston, MA, Nov. 5-8, 2000, sponsored by SPIE, 10 pp.
Zhang et al. “Development of Chemical Oxygen Demand On-Line Monitoring System Based on a Photoelectrochemical Degradation Principle” Environ. Sci. Technol., 40(7):2363-2368 (2006).
“U.S. Army Fitness Training Handbook” by the Department of the Army, 2003, The Lyons Press, p. 17.
“Warfighter Physiological and Environmental Monitoring: A Study for the U.S. Army Research Institute in Environmental Medicine and the Soldier Systems Center”, Massachusetts Institute of Technology Lincoln Laboratory, Final Report, Nov. 1, 2004, prepared for the U.S. Army under Air Force Contract F19628-00-C-0002; approved for public release.
Colligan, M. J. et al. in “The psychological effects of indoor air pollution”, Bulletin of the New York Academy of Medicine, vol. 57, No. 10, Dec. 1981, p. 1014-1026.
De Paula Santos, U. et al, in “Effects of air pollution on blood pressure and heart rate variability: a panel study of vehicular traffic controllers in the city of Sao Paulo, Brazil”, European Heart Journal (2005) 26, 193-200.
Ebert, T et al., “Influence of Hydration Status on Thermoregulation and Cycling Hill Climbing,” Med. Sci. Sport Exerc. vol. 39, No. 2, pp. 323-329, 2007.
Falkner et al, “Cardiovascular response to mental stress in normal adolescents with hypertensive parents. Hemodynamics and mental stress in adolescents,” Hypertension 1979, 1:23-30.
Fleming et al., “A Comparison of Signal Processing Techniques for the Extraction of Breathing Rate from the Photopethysmorgram,” World Academy of Science, Engineering and Technology, vol. 30, Oct. 2007, pp. 276-280.
Geladas et al., “Effect of cold air inhalation on core temperature in exercising subjects under stress,” The American Physiological Society, pp. 2381-2387, 1988.
Gold, D.R. et al. in “Ambient Pollution and Heart Rate Variability”, Circulation 2000, 101:1267-1273.
International Search Report and Written Opinion of the International Searching Authority, corresponding to PCT/US2012/0948079, dated Oct. 9, 2012.
International Search Report Corresponding to International Application No. PCT/US2012/022634, dated Aug. 22, 2012, 9 pages.
International Search Report corresponding to International Patent Application No. PCT/US2012/046446, dated Jan. 14, 2013, 3 pages.
Maomao et al., “Mobile Context-Aware Game for the Next Generation,” 2nd International Conference on Application and Development of Computer Games ADCOG 2003, p. 78-81.
Maughan et al., “Exercise, Heat, Hydration and the Brain,” Journal of the American College of Nutrition, vol. 26, No. 5, pp. 604S-612S, 2007.
Maughan, R.J., “Impact of mild dehydration on wellness and on exercise performance,” European Journal of Clinical Nutrition, 57, Suppl. 2, pp. S19-S23, 2003.
Mostardi, R., et al., “The effect of increased body temperature due to exercise on the heart rate and the maximal aerobic power,” Europ. J. Appl. Physiol, 33, pp. 237-245, 1974.
Nakajima et al., “Monitoring of heart and respiratory rates by photoplethyusmography using a digital filtering technique,” Med. Eng. Phys., vol. 18, No. 5, Jul. 1996, pp. 365-372.
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated Jul. 30, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/021936.
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated Aug. 26, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/021629.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Search Authority dated Sep. 16, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/024922.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Search Authority dated Sep. 27, 2010 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2010/025216.
Notification of Transmittal of The International Search Report and The Written Opinion of the International Searching Authority, or the Declaration corresponding to International Application No. PCT/US2013/070271; dated Feb. 26, 2014; 13 pages.
Shorten et al., “Acute effect of environmental temperature during exercise on subsequent energy intake in active men,” Am. J Clin. Nutr. 90, pp. 1215-1221, 2009.
Thompson, M.W., “Cardiovascular drift and critical core temperature: factors limiting endurance performance in the heat?” J. Exerc. Sci. Fit, vol. 4, No. 1, pp. 15-24, 2006.
Edmison et al., “E-Textile Based Automatic Activity Diary for Medical Annotation and Analysis,” Proc. BSN 2006 Int. Workshop Wearable Implantable Body Sensor Netw. (2006), pp. 131-145, Apr. 3-5, 2006.
Gibbs et al., “Reducing Motion Artifact in Wearable Bio-Sensors Using MEMS Accelerometers for Active Noise Cancellation,” 2005 American Control Conference, Jun. 8-10, 2005, Portland, OR, USA, pp. 1581-1586.
International Search Report corresponding to International Patent Application No. PCT/US2014/012909, dated May 13, 2014, 3 pages.
Notification Concerning Transmittal of International Preliminary Report on Patentability, PCT/US2014/012909, dated Jul. 28, 2015.
Wood et al., “Active Motion Artifact Reduction for Wearable Sensors Using Laguerre Expansion and Signal Separation,” Proceedings of the 2005 IEEE Engineering in Medicine and Biology, 27th Annual Conference. Shanghai, China, Sep. 1-4, 2005, pp. 3571-3574.
International Preliminary Report on Patentability, PCT/US2014/012940, dated Jun. 17, 2015, 23 pages.
International Search Report and Written Opinion of the International Searching Authority, corresponding to International Patent Application No. PCT/US2014/012940, dated Oct. 16, 2014, 13 pages.
Communication pursuant to Article 94(3) EPC, European Patent Application No. 13863449.8, dated Nov. 5, 2015, 7 pages.
Communication pursuant to Article 94(3) EPC, European Patent Application No. 14743615.8, dated Dec. 23, 2015, 7 pages.
Communication pursuant to Article 94(3) EPC, European Patent Application No. 14743839.4, dated Dec. 23, 2015, 6 pages.
European Search Report, EP Application No. 13863449.8, dated Oct. 19, 2015, 3 pages.
European Search Report, EP Application No. 14743615.8, dated Oct. 12, 2015, 3 pages.
European Search Report, EP Application No. 14743839.4, dated Oct. 12, 2015, 3 pages.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/014562, dated Oct. 28, 2015.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/042636, dated Oct. 29, 2015.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/042015, dated Oct. 29, 2015.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/042035, dated Oct. 29, 2015.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2015/046079, dated Dec. 29, 2015.
Communication pursuant to Article 94(3) EPC, European Patent Application No. 12820308.0, dated Feb. 3, 2016, 5 pages.
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated May 26, 2016 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2016/019126.
Notification of Transmittal of the International Search Report and Written Opinion of the International Search Authority dated May 26, 2016 by the Korean Intellectual Property Office for corresponding International Application No. PCT/US2016/019132.
Asada, et al., “Mobile Monitoring with Wearable Photoplethysmographic Biosensors,” IEEE Engineering in Medicine and Biology Magazine, May/Jun. 2003, pp. 28-40.
Bifulco et al., “Bluetooth Portable Device for Continuous ECG and Patient Motion Monitoring During Daily Life,” Medicon 2007, IFMBE Proceedings 16, 2007, pp. 369-372.
Brodersen et al., “In-Ear Acquisition of Vital Signs Discloses New Chances for Preventive Continuous Cardiovascular Monitoring,” 4th International Workshop on Wearable and Implantable Body Sensor Networks (BSN 2007), vol. 13 of the series IFMBE Proceedings, pp. 189-194.
Celka et al, “Motion Resistant Earphone Located Infrared based Heart Rate Measurement Device,” Proceedings of the Second IASTED International Conference on Biomedical Engineering, Feb. 16-18, 2004, Innsbruck, Austria, pp. 582-585.
Communication Pursuant to Article 94(3) EPC, EP 12 739 502.8, dated Jul. 19, 2016, 7 pages.
Communication Pursuant to Article 94(3) EPC, EP 14 743 615.8, dated Jul. 19, 2016, 7 pages.
Communication Pursuant to Article 94(3) EPC, EP 14 743 839.4, dated Jul. 20, 2016, 5 pages.
Comtois et al., “A Wearable Wireless Reflectance Pulse Oximeter for Remote Triage Applications,” 2006 IEEE, pp. 53-54.
Comtois, Gary, W., “Implementation of Accelerometer-Based Adaptive Noise Cancellation in a Wireless Wearable Pulse Oximeter Platform for Remote Physiological Monitoring and Triage,” Thesis, Worcester Polytechnic Institute, Aug. 31, 2007, 149 pages.
Duun et al., “A Novel Ring Shaped Photodiode for Reflectance Pulse Oximetry in Wireless Applications,” IEEE Sensors 2007 Conference, pp. 596-599.
Geun et al., “Measurement Site and Applied Pressure Consideration in Wrist Photoplethysmography,” The 23rd International Technical Conference on Circuits/Systems, Computers and Communications, 2008, pp. 1129-1132.
Gibbs et al., “Active motion artifact cancellation for wearable health monitoring sensors using collocated MEMS accelerometers,” Smart Structures and Materials, 2005: Sensors and Smart Structures Technologies for Civil, Mechanical, and Aerospace Systems, Proc. of SPIE, vol. 5765, pp. 811-819.
Haahr et al., “A Wearable “Electronic Patch” for Wireless Continuous Monitoring of Chronically Diseased Patients,” Proceedings of the 5th International Workshop on Wearable and Implantable Body Sensor Networks, in conjunction with The 5th International Summer School and Symposium on Medical Devices and Biosensors, The Chinese University of Hong Kong, HKSAR, China, Jun. 1-3, 2008, pp. 66-70.
Jiang, Honghui, “Motion-Artifact Resistant Design of Photoplethysmograph Ring Sensor for Driver Monitoring,” Thesis, Massachusetts Institute of Technology, Feb. 2004, 62 pages.
Kuzmina et al., “Compact multi-functional skin spectrometry set-up,” Advanced Optical Materials, Technologies, and Devices, Proc. of SPIE, vol. 6596, 2007, pp. 65960T-1 to 65960T-6.
Lee et al, “Respiratory Rate Detection Algorithms by Photoplethysmography Signal Processing,” 30th Annual International IEEE EMBS Conference, Vancouver, British Columbia, Canada, Aug. 20-24, 2008, pp. 1140-1143.
Lindberg et al., “Monitoring of respiratory and heart rates using a fibre-optic sensor,” Med Biol Eng Comput, Sep. 1992, vol. 30, No. 5, pp. 533-537.
Luprano, Jean, “Sensors and Parameter Extraction by Wearable Systems: Present Situation and Future,” pHealth 2008, May 21, 2008, 29 pages.
Lygouras et al., “Optical-Fiber Finger Photo-Plethysmograph Using Digital Techniques,” IEEE Sensors Journal, vol. 2, No. 1, Feb. 2002, pp. 20-25.
Maguire et al., “The Design and Clinical Use of a Reflective Brachial Photoplethysmograph,” Technical Report NUIM/SS/--/2002/04, Submitted Apr. 2002, Signals and Systems Research Group, National University of Ireland, Maynooth, Co. Kildare, Ireland, 13 pages.
Mendelson et al., “Measurement Site and Photodetector Size Considerations in Optimizing Power Consumption of a Wearable Reflectance Pulse Oximeter,” Proceedings of the 25th Annual International Conference of the IEEE EMBS, Cancun, Mexico, Sep. 17-21, 2003, pp. 3016-3019.
Mendelson et al., “Noninvasive Pulse Oximetry Utilizing Skin Reflectance Photoplethysmography,” IEEE Transactions on Biomedical Engineering, vol. 35, No. 10, Oct. 1988, pp. 798-805.
Poh et al., “Motion Tolerant Magnetic Earring Sensor and Wireless Earpiece for Wearable Photoplethysmography,” IEEE Transactions on Information Technology in Biomedicine, vol. 14, No. 3, May 2010, pp. 786-794.
Renevey et al., “Wrist-Located Pulse Detection Using IR Signals, Activity and Nonlinear Artifact Cancellation,” IEEE EMBS, 2001, 4 pages.
Rhee et al., “Artifact-Resistant Power-Efficient Design of Finger-Ring Plethysmographic Sensors,” IEEE Transactions on Biomedical Engineering, vol. 48, No. 7, Jul. 2001, pp. 795-805.
Shaltis, Phillip Andrew, Analysis and Validation of an Artifact Resistant Design for Oxygen Saturation Measurement Using Photo Plethysmographic Ring Sensors, Thesis, Massachusetts Institute of Technology, Jun. 2004, 103 pages.
Shin et al., “A Novel Headset with a Transmissive PPG Sensor for Heart Rate Measurement,” ICBME 2008, Proceedings 23, 2009, pp. 519-522.
Spigulis et al., “Wearable wireless photoplethysmography sensors,” Proc. of SPIE, vol. 6991,2008, pp. 69912O-1 to 69912O-7.
Takatani et al., “Optical Oximetry Sensors for Whole Blood and Tissue,” IEEE Engineering in Medicine and Biology, Jun./Jul. 1994, pp. 347-357.
Vogel et al., “A System for Assessing Motion Artifacts in the Signal of a Micro-Optic In-Ear Vital Signs Sensor,” 30th Annual International IEEE EMBS Conference, Vancouver, British Columbia, Canada, Aug. 20-24, 2008.
Vogel et al., “In-Ear Heart Rate Monitoring Using a Micro-Optic Reflective Sensor,” Proceedings of the 29th Annual International Conference of the IEEE EMBS Cite Internationale, Lyon, France, Aug. 23-26, 2007, pp. 1375-1378.
Wang et al., “Multichannel Reflective PPG Earpiece Sensor With Passive Motion Cancellation,” IEEE Transactions on Biomedical Circuits and Systems, vol. 1, No. 4, Dec. 2007, pp. 235-241.
Wang et al., “Reflective Photoplethysmograph Earpiece Sensor for Ubiquitous Heart Rate Monitoring,” 4th International Workshop on Wearable and Implantable Body Sensor Networks, 2007, vol. 13 of the series IFMBE Proceedings, pp. 179-183.
Wei et al. “A New Wristband Wearable Sensor Using Adaptive Reduction Filter to Reduce Motion Artifact,” Proceedings of the 5th International Conference on Information Technology and Application in Biomedicine, in conjunction with The 2nd International Symposium & Summer School on Biomedical and Health Engineering, Shenzhen, China, May 30-31, 2008, pp. 278-281.
Wood, Levi Benjamin, “Motion Artifact Reduction for Wearable Photoplethysmogram Sensors Using Micro Accelerometers and Laguerre Series Adaptive Filters,” Thesis, Massachusetts Institute of Technology, Jun. 2008, 74 pages.
Han et al., “Artifacts in wearable photoplethysmographs during daily life motions and their reduction with least mean square based active noise cancellation method,” Computers in Biology and Medicine, 42, 2012, pp. 387-393.
Extended European Search Report, EP Application No. 16164775.5 dated Sep. 13, 2016, 7 pages.
Notification of Transmittal of the International Search Report and the Written Opinion of the International Searching Authority, or the Declaration, PCT/US2016/041842, dated Oct. 21, 2016, 5 pages.
Notification of Transmittal of International Preliminary Reporton Patentability, PCT/US2015/041562, dated Oct. 20, 2016, 14 pages.
Notification of Transmittal of International Preliminary Reporton Patentability, PCT/US2015/042636, dated Oct. 20, 2016, 7 pages.
Notification of Transmittal of International Preliminary Report on Patentability, PCT/US2015/042015, dated Oct. 20, 2016, 10 pages.
Notification of Transmittal of International Preliminary Report on Patentability, PCT/US2015/042035, dated Oct. 20, 2016, 8 pages.
Notification of Transmittal of International Preliminary Reporton Patentability, PCT/US2015/046079, dated Oct. 20, 2016, 10 pages.
Webster, J. G. Design of Pulse Oximeters. IOP Publishing Ltd., 1997, Cover page, pp. i-xvi, pp. 34-159.
Communication with Supplementary European Search Report, European Application No. 15830336.2, dated Jun. 7, 2017, 8 pp.
Comtois et al., “A Comparative Evaluation of Adaptive Noise Cancellation Algorithms for Minimizing Motion Artifacts in a Forehead-Mounted Wearable Pulse Oximeter”, Proceedings of the 29th Annual International Conference of the IEEE EMBS, Lyon, France, Aug. 23-26, 2007, pp. 1528-1531.
Han et al. “Development of a wearable health monitoring device with motion artifact reduced algorithm” International Conference on Control, Automation and Systems 2007 (ICCAS. 2007), Seoul, Korea, Oct. 17-20, 2007, pp. 1581-1584.
Lee et al., “A Mobile Care System With Alert Mechanism”, IEEE Transactions on Information Technology in Biomedicine, vol. 11, No. 5, Sep. 2007, pp. 507-517.
Related Publications (1)
Number Date Country
20180220906 A1 Aug 2018 US
Provisional Applications (3)
Number Date Country
60905761 Mar 2007 US
60876128 Dec 2006 US
60875606 Dec 2006 US
Divisions (1)
Number Date Country
Parent 14328107 Jul 2014 US
Child 15949331 US
Continuations (2)
Number Date Country
Parent 14063669 Oct 2013 US
Child 14328107 US
Parent 11811844 Jun 2007 US
Child 14063669 US