Monitoring system for providing both visual and non-visual data

Information

  • Patent Grant
  • 11463660
  • Patent Number
    11,463,660
  • Date Filed
    Friday, May 3, 2019
    5 years ago
  • Date Issued
    Tuesday, October 4, 2022
    a year ago
Abstract
Monitoring systems are provided that include both visual and non-visual data to monitor the well-being of individuals, such as infants and patients.
Description
FIELD OF THE DISCLOSURE

The present disclosure relates to systems that include both visual and non-visual data to monitor the well-being of individuals, such as infants and patients.


BACKGROUND

Smart devices and systems for monitoring the care of infants, ageing adults, and patients are becoming commonplace. Camera and audio based baby monitors can be used to determine if an infant is sleeping or awake, content or upset, and the general nature of the infant via visual observation. However these monitors generally do not yield quantitative data regarding sleep, activity level and patterns thereof. Other monitoring devices associated with a diaper, such as those described in U.S. Pat. No. 8,628,506, can be used to determine the location, body position, and physical activity of the diaper wearer. In addition to the '506 patent being silent regarding a camera, a skilled artisan reviewing this patent can likely conclude visual data is unneeded with the scope of non-visual data available by the disclosed devices. However, a caregiver, particularly a parent of an infant, can have an emotional need to be able to see his/her infant with their own eyes and/or hear the infant with their own ears even if a device associated with the infant's diaper or clothing provides non-visual data in all of the categories the parent is interested in.


Monitoring systems of the present disclosure have the capability to provide both visual and non-visual data about the nature of a monitored individual.


SUMMARY

Monitoring systems of the present disclosure have the capability to provide visual and non-visual data about the nature of a monitored individual. Exemplary monitoring systems can comprise a wearable sensor device to generate non-visual data and a camera device to generate visual data. The wearable sensor device can be associated with a disposable absorbent article or clothing that is worn by an infant or patient being monitored.


In accordance with one form, a monitoring system is provided comprising a first monitoring device capable of providing non-visual data regarding sleep or motion-related characteristics of a monitored individual; a second monitoring device capable of providing visual data regarding sleep or motion-related characteristics of the monitored individual; and a software application for displaying both the non-visual data or aspects thereof, and the visual data or aspects thereof, to a caregiver of the monitored individual.


In accordance with another form, a monitoring system is provided comprising a multi-use wearable sensor device for temporary association with an absorbent article, and a camera device capable of capturing visual information regarding the monitored individual. The multi-use wearable sensor device comprises a first sensor capable of capturing non-visual data/information regarding a monitored individual and a second sensor capable of capturing a condition of the absorbent article.





BRIEF DESCRIPTION OF THE DRAWINGS

The above-mentioned and other features and advantages of the present disclosure, and the manner of attaining them, will become more apparent and the disclosure itself will be better understood by reference to the following description of example forms of the disclosure taken in conjunction with the accompanying drawings, wherein:



FIG. 1 is a schematic showing an exemplary monitoring system comprising a wearable sensor device and a camera device. FIG. 1 also shows a peripheral device that can be used to display and monitor data and other information collected by the wearable sensor device and/or camera device.



FIG. 2 is a schematic showing a wearable sensor device attached to an exterior of a diaper.



FIG. 3 is a front view of an exemplary camera device suitable for use in monitoring systems provided herein.



FIG. 4 is side view of an exemplary camera device that includes a camera and a camera docking station.



FIG. 5 is a perspective view of an exemplary camera device that is attached to a first docking station along with a second docking station for placement in a second location.



FIG. 6 is a diagram of an exemplary monitoring system comprising a data gathering module, a data processing module, and a communication module.



FIG. 7 is a flow diagram showing one operational scheme for monitoring systems of the present disclosure.



FIG. 8 is a partial view of an exemplary wearable sensor device of the present disclosure.



FIG. 9 is a caregiver-facing side of a wearable sensor device of the present disclosure.



FIG. 10 is an exemplary kit of the present disclosure, comprising a plurality of disposable absorbent articles, two wearable sensor devices, a camera, and two camera docking stations.





DETAILED DESCRIPTION

Various non-limiting aspects of the present disclosure will now be described to provide an overall understanding of the principles of the structure, function, manufacture, and/or use of the subject matter disclosed herein. In the following detailed description, reference is made to the accompanying figures, which form a part hereof. In the figures, similar symbols can identify similar elements, unless context dictates otherwise. It will be readily understood that the aspects of the present disclosure, as generally described herein, and illustrated in the figures, can be arranged, substituted, combined, separated, and designed in a wide variety of different configurations.


Monitoring systems of the present disclosure have the capability to provide both visual and non-visual data about the nature of a monitored individual. With reference to FIG. 1, an exemplary monitoring system 10 is shown that comprises a wearable sensor device 20 to generate non-visual data and a camera device 30 to generate visual data and optionally audio. Wearable sensor device 20 can be associated with a disposable absorbent article (see, e.g., FIG. 2) or clothing that is worn by an infant or patient being monitored. Wearable sensor device 20 can be a single-use sensor or a multi-use sensor that can be removed from clothing or a soiled disposable absorbent article and associated with a freshly-donned article. A multi-component wearable sensor device is also contemplated, wherein a first sensor component (e.g., capacitor) forms a part of the manufactured article and a second, multi-use sensor component (e.g., power source and data transmitter) connects with the first sensor component to create sensor functionality.


A peripheral device 40 in the form of a smart phone is shown in FIG. 1, as an example of a device that can be used by a caregiver to observe and manage the non-visual and/or visual data or aspects thereof that are provided by monitoring systems of the present disclosure. The non-visual data transmitted from wearable sensor device 20 can be communicated to camera device 30 and thereafter communicated to a peripheral device, or alternatively, be communicated to the peripheral device bypassing camera device 30. Camera device 30 can also be configured to communicate non-visual data from wearable sensor device 20 in the absence of communicating visual data obtained by camera device 30.


Monitoring systems of the present disclosure can optionally include additional sensors beyond wearable sensor device 20. For example, an environmental sensor to measure environmental conditions where an infant or patient is located can be used. A list of non-limiting environmental sensors includes a temperature sensor, a relative humidity sensor, a carbon monoxide sensor, a VOC (volatile organic compound) sensor, a smoke sensor, a motion sensor, and combinations thereof. These additional sensors can be separate from wearable sensor device 20 and camera device 30, or alternately be incorporated with one or both of them. By way of example only, FIG. 3 shows a camera device 30′ that includes a sensor port 32 that allows a temperature sensor and/or relative humidity sensor disposed within camera device 30′ to measure such environmental conditions.


Camera devices forming part of the disclosed monitoring systems can take many forms and can include various functionality, including an ability to capture video and still visual images, contain night vision technology to capture visual images in a dark or semi-dark environment, contain noise and/or motion activation features, and ability to transmit current and historical visual information. An exemplary camera device 50, shown in FIGS. 4 and 5, includes a camera head 52 pivotally connected to an arm 54 that is attached to a support member in the form of a base 56. Support members of suitable camera devices can take many forms and can include a variety of features for securing the camera device to its intended structure or location. Having a pivotable or otherwise moveable camera head can provide numerous benefits, including, for example, being able to capture visual data from different fields of view without moving the camera base position. In some forms, the gap between camera head 52 and arm 54 or base 56 can be between about 3 mm and 16.5 mm to help mitigate pinching and entrapment. As can be seen in FIG. 4, camera head 52 can also pivot downward to face base 56 for privacy and to prevent capturing unintentional visual data.


Exemplary camera device 50 further includes a docking station 58 that is releasably engageable with base 56. A latch 60 is employed to positively secure base 56 to docking station 58. Other means, for example, magnets, can equally be used for making the secure connection between base 56 and docking station 58. The monitoring systems can include multiple docking stations, as shown in FIG. 5, to enable capturing visual data from multiple locations with a single camera. The multiple docking stations can be similar or dissimilar to one another. For example, one of the multiple docking stations can be configured (e.g., with brackets to secure the docking station power cord to a wall) for a more permanent location such as a nursery, while a second docking station is configured for mobility to be taken to different locations within or outside of one's home. The skilled artisan should readily appreciate that multiple standalone camera devices can be included with monitoring systems disclosed herein rather than having a single camera and multiple docking stations.


Besides optionally comprising additional sensors, the camera devices of the monitoring systems herein can include light sources and audio capability. The light sources can emit constant or changing light effects. The light source can, for example, emit light outside of the visible blue light spectrum (e.g., wavelengths in the range of 300 nm to 700 nm, or 400 nm to 500 nm). Blue light hazard function weight radiance (LB) can be less than 0.2. The audio can be one-way or two-way. The light sources can avoid blinking and flashing LEDs with frequency between 3 Hz and 60 Hz to help mitigate seizures in photosensitive individuals (both monitored individuals and caregivers). The audio capability can permit communication between a caregiver and a monitored infant or patient, including from the infant/patient to the caregiver. The audio capability can also provide caregiver communications, music or other soothing sounds.


The camera devices of the monitoring systems herein can also include a variety of safety features. A mount of the camera device and/or secure positioning within a docking station should be able to sustain certain loading before the camera device or camera within the docking station is dislodged. For example, the placed positioning can withstand three times the device or component weight applied downwardly for 1 minute and/or 50 newtons applied sideways for 1 minute (after stress relief testing) according to IEC 60065-14 wall mounting test (section 19.7). The camera device with or without a separable docking station can be sufficiently light to help mitigate harm to a monitored individual or caregiver should it accidently fall on them. Along these lines, the camera device (with or without a docking station) can weigh less than 500, 400, 300, 200, 100, or even 75 grams.


Power cords that are extremely flexible can more easily become wrapped around an undesirable object such as a monitored individual. To help prevent this, the power cord of the camera devices can have a safe release mechanism or easy cord pull (e.g., under 0.65 pound force) which can help mitigate strangulation concerns. Increasing the bending stiffness and/or its cross-sectional areas may also mitigate strangulation hazards.


The wearable sensor device can be a single-use device, or a multi-use device. The wearable sensor device can be attachable to skin (via hydrogel or bio-adhesive material, for example), to a disposable article such as an absorbent article, and/or to clothing worn by a monitored individual. Various attachment mechanisms can be employed for attaching the wearable sensor device to an article of clothing or disposable absorbent article such as a diaper, pant, pad or brief. For example, hook and loop fastening mechanisms, magnets, adhesives, thermal bonds, and male and female mating fasteners such as snaps and buttons. Receiving features, such as pockets, recesses, and voids can also be employed that essentially hold the wearable sensor device with or without attachment features. In yet another form, an auxiliary article can be used to integrate the wearable sensor device with a disposable absorbent article. The auxiliary article can be in the form of a pant-like reusable garment designed to fit over a disposable absorbent article.


In one form, the wearable sensor device is adapted for attachment to an outer cover of a disposable absorbent article. Hook and loop features can be used with this attachment approach. For example, a strip of hook material can be affixed to one surface of the sensor housing, where the hooks can engage directly with material used for the outer cover or with an added strip of “loop” material.


The wearable sensor device can include a sensor housing that generally protects sensors and other electronic components disposed therein, as well as inhibiting unwanted contact of the same with an infant, patient, or caregiver. The housing can be made from a variety of materials, both flexible and rigid, examples of which include thermoplastic polymers, thermoplastic elastomers, silicone, Tecaform, Tecanant, thermoplastic copolyester (TPC) and combinations thereof. Other materials can also be employed for the housing so long as it is generally regarded as safe for human contact and does not cause irritation or other unwanted health effects. The materials can include bio-compatible, medical grade, and non-cytotoxic materials. Inclusion of a bittering agent or other approaches can optionally be used to discourage placement of the sensor device in one's mouth or otherwise tampering with the wearable sensor device.


The wearable sensor device comprises a plurality of electronic component disposed on and/or within the housing. Typically, the electronic components include at least one sensor, a transmitter, and a power source (e.g., a disposable battery or a rechargeable battery). The number and type of sensors employed by the sensor device are chosen based on the application of the monitoring systems disclosed herein.


The monitoring systems are capable of capturing and communicating a combination of visual data and non-visual data to caregiver. With connection back to the background section above, one exemplary monitoring system of the present disclosure includes the combination of a wearable sensor device comprising a motion sensor and a camera device. The motion sensor can provide non-visual sleep-related data including, for example, time of sleep, sleep duration, sleep position, sleep position changes, and activity-related data including, for example, tummy time, level of activity, and nursing/feeding. Suitable motion sensors include accelerometers, inertial measurement units (IMUs), gyroscopes, and magnetometers. As noted in the background, notwithstanding the benefit of obtaining the non-visual sleep-related data, a caregiver wants/needs to see a monitored individual to understand visual sleep (or pre- and/or post-sleep) aspects of a monitored individual. The camera device can provide this visual confirmation desired by the caregiver. Having both data sources available additionally provides comfort to a caregiver when the camera device is unable to capture information regarding the monitored individual. For example, the camera device is powered off or the monitored individual is not within the field of view of the camera. In this scenario the wearable sensor device can provide data and information to the caregiver until the camera device is able to capture the monitored individual.


Besides obtaining sleep-related data and information, monitoring systems provided herein are capable of obtaining both visual and non-visual data and information related to other activities that contain a motion aspect, including, for example, eating, drinking, breastfeeding, walking, crawling, and tummy-time. The non-visual data and information can be determined by a wearable sensor device or a remote sensor device (unattached to a monitored individual or anything worn by the individual) that is capable of measuring motion and a camera device.


Some monitoring systems of the present disclosure can include a wearable sensor device comprising two or more sensors and a camera device. The wearable sensor device is associated with a disposable absorbent article and comprises a first sensor that is capable of capturing non-visual data and information regarding a monitored individual, and a second sensor that is capable of capturing a condition (e.g., clean or soiled with body exudates) of the disposable absorbent article. The first sensor associated with the wearable sensor device can be a motion sensor as described above for capturing information about sleep or activity aspects. The second sensor can be a wetness sensor or bowel movement (BM) sensor for detecting urine or feces within the disposable absorbent article, and/or track absorbent article usage. The wearable sensor device can include both a wetness sensor and a BM sensor in some forms. In one form, the wearable sensor device is free of audio capturing or communicating components.


Wetness sensors for detecting the presence of urine or other bodily fluid can include optical sensors, color sensors, and electrical sensors that comprise a resistance, capacitance, inductance or continuity sensitive indicator. A resistance sensitive indicator can be provided, for example, by providing two electrical conductors disposed at a given spatial distance relative to each other. A VOC sensor is one suitable type of a BM sensor. The VOC sensors can be of MOS-type (metal oxide). BM sensors can be capable of acting as an electronic nose to detect chemical signatures of organic materials associated with body exudates, including, for example, skatole, mercaptans, amines, volatile fatty acids, acetone, methyl acetate, and methanol. BM sensors can also include an optical or color sensor to detect the presence of feces in the article. Along these lines, multiple optical or color sensors can be used to detect both urine and feces, based either on their inherent colors or based on use of an indicator that changes color in the presence of urine and/or feces. For example, the following enzymes associated with body exudates can trigger an optical change in an included indicator that can be sensed by an optical or color sensor: urease, trypsin, chemotrypsin, LAP, lipase, amilase, and urease.


With reference to FIG. 6, monitoring systems S1 of the present disclosure can include three elements: a data gathering module 100, a data processing module 102, and a communication module 103. The data gathering module can include a camera device 202/205 and a wearable sensor device 200 and that can be associated with a disposable absorbent article or article of clothing worn by an infant or patient. Besides the wearable sensor device and camera device, the data gathering module can comprise other sensor devices or equipment in proximity to the person wearing the sensor device. For example, the data gathering module can comprise an environmental sensor for sensing smoke, carbon monoxide, VOC's, temperature, relative humidity; a motion sensor, an audio recorder, and the like.


The data processing module can comprise data transmission, data storage, data interpretation, and/or data manipulation to transform the data from the data gathering module into consumer understandable information related to the wellbeing of an individual, including, for example, feeding, sleeping, and/or voiding. And the communication module comprises a software application for communicating (e.g., displaying) the data and information from the wearable sensor device and the camera device.



FIG. 7 is an exemplary schematic illustrating how monitoring systems of the present disclosure can operate. Two example operation modes are shown; one where the wearable sensor device 200 is worn by an infant/patient that is located within a dwelling or care center 202 with wireless communication (e.g., Bluetooth low energy, 15.4, ad hoc mesh networks, and the like) conducted between wearable sensor device 200 and a hub 202 comprising a camera device 205. And another where the infant/patient is located remotely (e.g., in a car or stroller) with wireless communication conducted between wearable sensor device 200 and a smart peripheral device 204 (e.g., phone). Data and information from wearable sensor device 200 is communicated via a router 203 or smart device 204 to a system frontend 206 for transforming the data and information to consumer usable information provided via a software application.



FIGS. 8 and 9 illustrate aspects of one exemplary wearable sensor device 300 comprising a plurality of sensors and other electronic components. FIG. 8 shows the wearable sensor device in a manufactured form before its final configuration and encasing with a sensor housing 310 (shown in FIG. 10). With reference to FIG. 8, multiple electrically-connected printed circuit boards 312, 313, 314, and 315 are employed. Printed circuit board 312 comprises an optical sensor 320, an absorbent article-facing light source 322, a power management component 324, a flash memory component 326, an optical sensor frontend 328, a processor and transmitter component 330, and an antenna 332. Printed circuit board 313 comprises a motion sensor (e.g., accelerometer) 334 and a power source (e.g., coin cell battery) 336. Printed circuit board 314 comprises a BM sensor (e.g., VOC sensor) 338, a temperature and relative humidity sensor 340, a second optical sensor 342, and a second absorbent article-facing light source 344. And printed circuit board 315 comprises a consumer-engageable button 346 for activating or otherwise operating sensor device 300, and a caregiver-facing light source 348 to indicate an operational aspect of sensor device 300. Button 346 or similar engageable feature can be used for multiple tasks. For example, button 346 can be initially activated for “waking” the wearable sensor device up from a power-save mode and/or manually activated by a caregiver upon changing an absorbent article if a wearer of the article has had a bowel movement. Acknowledgement of a bowel movement via action of button 346 or similar engageable feature can be communicated by the wearable sensor device to a communication module for tracking timing, frequency, or other aspects of a wearer's bowel movement history. One skilled in the art would appreciate that a single circuit board can be employed in a sensor device, as well as other numbers of circuit boards beyond what is shown in FIG. 8.


In one form, the wearable sensor device comprises a power source in the form of a battery, a transmitter, multiple optical sensors (e.g., a color sensor), multiple light sources (e.g., an LED), and an accelerometer. The wearable sensor device is attached to an absorbent article comprising a wetness indicator, as described above, such that the light source can direct light onto the wetness indicator. The wetness indicator changes appearance (e.g., changes color) when a wearer urinates into the absorbent article. The optical sensor measures the reflected light from the wetness indicator to sense when a urination event occurs. Multiple pairs of absorbent article-facing light sources and optical sensors can be employed to sense changes of a wetness indicator at different points along the indicator to confirm a urination event has occurred, or predict the volume of urine and/or number of urination events that occurred. A signal from the optical sensor can then be transmitted to the data processing module. The accelerometer is employed to track data associated with sleep and awake times. The awake data can include awake feeding motion data and awake non-feeding motion. The accelerometer is capable of sensing breastfeeding times and provide feeding information for one's right breast and left breast. The accelerometer can also be configured to sense bottle feeding aspects.


The data processing module can comprise data transmission, data storage, data interpretation, data filtering, and/or data manipulation to transform the data from the data gathering module into consumer understandable information related to the wellbeing of an individual, including, for example, feeding, sleeping, and/or voiding. The data processing module can include algorithms to parse/filter the received data. Data processing can be accomplished by one or more devices and in the same or different locations. For example, the wearable sensor device can optionally employ a memory device to temporarily store data. One reason for temporary storage of data is when communication between the wearable sensor device and a remote data processing module component and/or the information communication module is unavailable.


The wearable sensor device can also optionally employ a data processor for processing raw data from one or more sensors associated with the wearable sensor device prior to transmitting data/information based on the raw data. This can reduce the volume of data/information transmitted from the wearable sensor device, and thereby reduce the amount of power required and accompanying electromagnetic radiation emission.


The communication module comprises a software application operable on a computer device to display information related to the data obtained by the data gathering module, including data transformed via the data processing module. The computer device can be a smart phone, as is shown in FIG. 1, but other computer devices, such as a laptop, tablet, digital assistant (ALEXA and GOOGLE HOME, for example) can be used to communicate information to one or more caregivers. Monitoring systems of the present disclosure can provide both visual and non-visual data and information regarding a monitored individual. Software applications are typically employed to display and otherwise communicate this data and information. While separate software applications can be employed to view and/or manage the visual and non-visual components, respectively, it is preferred for a single software application to be utilized to both provide visual data of the monitored individual and provide the non-visual data and information arising from a wearable sensor device and a camera device. Data and information from the wearable sensor device and/or camera device can be encrypted in the event someone other than the caregiver gains access to the data and information.


Wearable sensor devices and monitoring systems including the same can form a part of consumer purchasable kit. One exemplary kit includes two or more wearable sensor devices as described herein, a camera device, a plurality of absorbent articles that can accept the wearable sensor devices, and access to a software application for viewing data and information flowing from the wearable sensor device and camera device. A subscription can be offered to consumers that provides delivery of additional absorbent articles, wearable sensor devices, and/or continued access and operation of the software application. For example, a subscription can include an automatic delivery of a number of absorbent articles every two weeks along with a code or other mechanism for continued operation of the software application. The subscription can work with an affirmative action request by a consumer or as an automatic delivery order that delivers products on a set re-occurring schedule until the schedule reaches a predetermined endpoint, or is altered or discontinued by the consumer. Similar to the subscription example above, packages of absorbent articles can be sold in brick and mortar locations wherein the packages contain a code for operation of the software application to view data and information received from a wearable sensor device according to the present disclosure.



FIG. 10 shows an exemplary kit 400 comprising a plurality of absorbent articles 402, two wearable sensor devices 404, a camera 406, and two camera docking stations 408.


While the discussion has focused on infants and patients, systems of the present invention are also applicable for elderly care. The sensors associated with such systems can be capable of association with an elderly's skin, durable undergarments, disposable absorbent articles, bed materials, bed pads, and/or clothing articles.


This application claims the benefit of U.S. Provisional Application No. 62/773,401, filed on Nov. 30, 2018, 62/723,179, filed on Aug. 27, 2018 and 62/666,989, filed on May 4, 2018, the entireties of which are all incorporated by reference herein.


The dimensions and values disclosed herein are not to be understood as being strictly limited to the exact numerical values recited. Instead, unless otherwise specified, each such dimension is intended to mean both the recited value and a functionally equivalent range surrounding that value. For example, a dimension disclosed as “40 mm” is intended to mean “about 40 mm.”


Every document cited herein, including any cross referenced or related patent or application and any patent application or patent to which this application claims priority or benefit thereof, is hereby incorporated herein by reference in its entirety unless expressly excluded or otherwise limited. The citation of any document is not an admission that it is prior art with respect to any disclosure disclosed or claimed herein or that it alone, or in any combination with any other reference or references, teaches, suggests or discloses any such disclosure. Further, to the extent that any meaning or definition of a term in this document conflicts with any meaning or definition of the same term in a document incorporated by reference, the meaning or definition assigned to that term in this document shall govern.


While particular forms of the present disclosure have been illustrated and described, it would be obvious to those skilled in the art that various other changes and modifications can be made without departing from the spirit and scope of the disclosure. It should be understood that other forms can include more or less of each element shown in a given figure. Further, some of the illustrated elements can be combined or omitted. Yet further, an exemplary form can include elements that are not illustrated in the figures. The various aspects and forms disclosed herein are for purposes of illustration and are not intended to be limiting, with the true scope and spirit being indicated by the following claims.

Claims
  • 1. A monitoring system, comprising: a. a first camera device or camera device component for enabling visual data feed from a first location;b. a second camera device or camera device component for enabling visual data feed from a second location; andc. a wearable sensor device,wherein at least one of the first or the second camera device or camera device component comprises one or more environmental sensors, the one or more environmental sensors comprising a temperature sensor and a relative humidity sensor.
  • 2. The monitoring system of claim 1, wherein the first camera device or camera device component comprises a camera, and the second camera device or camera component comprises a docking station capable of releasable engagement with the camera.
  • 3. The monitoring system of claim 1, wherein the first camera device or camera device component comprises a first docking station for releasable engagement with a camera, and the second camera device or camera component comprises a second docking station capable of releasable engagement with the camera.
  • 4. The monitoring system of claim 1, wherein the first camera device or camera device component comprises a first standalone camera, and the second camera device or camera component comprises a second standalone camera.
  • 5. The monitoring system of claim 1, wherein the wearable sensor device comprises an accelerometer.
  • 6. The monitoring system of claim 1, wherein the wearable sensor device comprises a wetness sensor.
  • 7. A method for monitoring an individual, the method comprising: a. obtaining visual data of a monitored individual from a first camera device at a first location;b. obtaining visual data of the monitored individual from a second camera device at a second location; andc. obtaining non-audiovisual data of the monitored individual from a wearable sensor device associated with an absorbent article worn by the monitored individual;wherein at least one of the first camera device or the second camera device comprises one or more environmental sensors, the one or more environmental sensors comprising a temperature sensor and a relative humidity sensor.
  • 8. The method of claim 7, further comprising obtaining audio from the monitored individual via at least one of the first or the second camera device.
  • 9. The method of claim 7, further comprising timewise aligning the obtained visual data and the obtained non-audiovisual data.
  • 10. The method of claim 7, further comprising: d. assigning a first timestamp to the obtained visual data;e. assigning a second timestamp to the obtained non-audiovisual data; andf. correlating the first timestamp and the second timestamp.
  • 11. The method of claim 7, further comprising displaying at least some of the visual data from at least one of the first or the second camera device simultaneously with at least some of the non-audiovisual data to a caregiver of the monitored individual.
US Referenced Citations (435)
Number Name Date Kind
3848594 Buell Nov 1974 A
3860003 Buell Jan 1975 A
3911173 Sprague Oct 1975 A
4022210 Glassman May 1977 A
4265245 Glassman May 1981 A
4286331 Anderson Aug 1981 A
4515595 Kievit et al. May 1985 A
4554662 Suzuki Nov 1985 A
4573986 Minetola et al. Mar 1986 A
4662875 Hirotsu et al. May 1987 A
4681793 Linman et al. Jul 1987 A
4695278 Lawson Sep 1987 A
4699622 Toussant et al. Oct 1987 A
4710189 Lash Dec 1987 A
4785996 Ziecker et al. Nov 1988 A
4795454 Dragoo Jan 1989 A
4808178 Aziz et al. Feb 1989 A
4842666 Werenicz Jun 1989 A
4846815 Scripps Jul 1989 A
4894060 Nestegard Jan 1990 A
4908803 Aziz et al. Mar 1990 A
4940464 Van Gompel et al. Jul 1990 A
4946527 Battrell Aug 1990 A
4963140 Robertson et al. Oct 1990 A
4977906 Di Scipio Dec 1990 A
5036859 Brown Aug 1991 A
5137537 Herron et al. Aug 1992 A
5151092 Buell et al. Sep 1992 A
5221274 Buell et al. Jun 1993 A
5242436 Weil et al. Sep 1993 A
5264830 Kline et al. Nov 1993 A
5354289 Mitchell et al. Oct 1994 A
5415649 Watanabe May 1995 A
5433715 Tanzer et al. Jul 1995 A
5469145 Johnson Nov 1995 A
5499978 Buell et al. Mar 1996 A
5507736 Clear et al. Apr 1996 A
5554145 Roe et al. Sep 1996 A
5569234 Buell et al. Oct 1996 A
5571096 Dobrin et al. Nov 1996 A
5580411 Nease et al. Dec 1996 A
5590152 Nakajima et al. Dec 1996 A
5607414 Richards et al. Mar 1997 A
5628097 Benson et al. May 1997 A
5700254 McDowall et al. Dec 1997 A
5709222 Davallou Jan 1998 A
5714156 Schmidt et al. Feb 1998 A
5817087 Takabayashi Oct 1998 A
5838240 Johnson Nov 1998 A
5865823 Curro Feb 1999 A
5902222 Wessman May 1999 A
5938648 LaVon et al. Aug 1999 A
5959535 Remsburg Sep 1999 A
6004306 Robles et al. Dec 1999 A
6160198 Roe et al. Mar 2000 A
6093869 Roe et al. Jul 2000 A
6121509 Ashraf et al. Sep 2000 A
6179820 Fernfors Jan 2001 B1
6203496 Gael et al. Mar 2001 B1
6246330 Nielsen Jun 2001 B1
6264643 Toyoda Jul 2001 B1
6306122 Narawa Oct 2001 B1
6372951 Ovanesyan et al. Apr 2002 B1
6384296 Roe et al. May 2002 B1
6432098 Kline et al. Aug 2002 B1
6501002 Roe et al. Dec 2002 B1
6534149 Daley et al. Mar 2003 B1
6583722 Jeutter Jun 2003 B2
6603403 Jeutter et al. Aug 2003 B2
6609068 Cranley Aug 2003 B2
6617488 Springer et al. Sep 2003 B1
6632504 Gillespie et al. Oct 2003 B1
6645190 Olson et al. Nov 2003 B1
6645569 Cramer et al. Nov 2003 B2
6761711 Fletcher et al. Jul 2004 B1
6817994 Popp et al. Nov 2004 B2
6840928 Datta et al. Jan 2005 B2
6849067 Fletcher et al. Feb 2005 B2
6863933 Cramer et al. Mar 2005 B2
6893426 Popp et al. May 2005 B1
6946585 London Brown Sep 2005 B2
6953452 Popp et al. Oct 2005 B2
6969377 Koele et al. Nov 2005 B2
7002054 Allen et al. Feb 2006 B2
7049969 Tamai May 2006 B2
7112621 Rohrbaugh et al. Sep 2006 B2
7145053 Emenike Dec 2006 B1
7156833 Courure-Dorschner et al. Jan 2007 B2
7174774 Pawar Feb 2007 B2
7201744 Van Gompel et al. Apr 2007 B2
7241627 Wilhelm et al. Jul 2007 B2
7250547 Hofmeister et al. Jul 2007 B1
7295125 Gabriel Nov 2007 B2
7355090 Alex et al. Apr 2008 B2
7394391 Long Jul 2008 B2
7410479 Hoshino Aug 2008 B2
7449614 Alex Nov 2008 B2
7477156 Long et al. Jan 2009 B2
7489252 Long et al. Feb 2009 B2
7497851 Koele et al. Mar 2009 B2
7498478 Long et al. Mar 2009 B2
7504550 Tippey et al. Mar 2009 B2
7524195 Alex et al. Apr 2009 B2
7527615 Roe May 2009 B2
7537832 Carlucci et al. May 2009 B2
7569039 Matsuda Aug 2009 B2
7595734 Long et al. Sep 2009 B2
7642396 Alex et al. Jan 2010 B2
7649125 Ales et al. Jan 2010 B2
7659815 Cohen et al. Feb 2010 B2
7667806 Ales et al. Feb 2010 B2
7682349 Popp et al. Mar 2010 B2
7700820 Tippey et al. Apr 2010 B2
7700821 Alex et al. Apr 2010 B2
7737322 Alex et al. Jun 2010 B2
7744579 Langdon Jun 2010 B2
7753691 Ales et al. Jul 2010 B2
7760101 Alex et al. Jul 2010 B2
7786341 Schneider et al. Aug 2010 B2
7789869 Berland et al. Sep 2010 B2
7803319 Yang et al. Sep 2010 B2
7812731 Benza et al. Oct 2010 B2
7834235 Long et al. Nov 2010 B2
7835925 Roe et al. Nov 2010 B2
7846383 Song Dec 2010 B2
7850470 Ales et al. Dec 2010 B2
7855653 Rondoni et al. Dec 2010 B2
7862550 Koele et al. Jan 2011 B2
7879392 Wenzel et al. Feb 2011 B2
7956754 Long Apr 2011 B2
7946869 Ales et al. May 2011 B2
7973210 Long et al. Jul 2011 B2
7977529 Berman et al. Jul 2011 B2
8007485 Popp et al. Aug 2011 B2
8044258 Hietpas Oct 2011 B2
8053624 Nhan et al. Nov 2011 B2
8053625 Nhan et al. Nov 2011 B2
8057454 Long et al. Nov 2011 B2
8058194 Nhan et al. Nov 2011 B2
8101813 Ales et al. Jan 2012 B2
8111165 Ortega et al. Feb 2012 B2
8115643 Wada et al. Feb 2012 B2
8172982 Ales et al. May 2012 B2
8173380 Yang et al. May 2012 B2
8183876 Wada et al. May 2012 B2
8196270 Mandzsu Jun 2012 B2
8196809 Thorstensson Jun 2012 B2
8207394 Feldkamp et al. Jun 2012 B2
8215973 Ales et al. Jul 2012 B2
8222476 Song et al. Jul 2012 B2
8237572 Clement et al. Aug 2012 B2
8248249 Clement et al. Aug 2012 B2
8264362 Ales et al. Sep 2012 B2
8274393 Ales et al. Sep 2012 B2
8299317 Tippey et al. Oct 2012 B2
8304598 Masbacher et al. Nov 2012 B2
8314284 Novello Nov 2012 B1
8334226 Nhan et al. Dec 2012 B2
8334425 Ales et al. Dec 2012 B2
8338659 Collins et al. Dec 2012 B2
8350694 Trundle Jan 2013 B1
8361048 Kuen et al. Jan 2013 B2
8372052 Popp et al. Feb 2013 B2
8372242 Ales et al. Feb 2013 B2
8372766 Nhan et al. Feb 2013 B2
8378167 Allen et al. Feb 2013 B2
8381536 Nhan et al. Feb 2013 B2
8384378 Feldkamp et al. Feb 2013 B2
8395014 Helmer et al. Mar 2013 B2
8416088 Ortega et al. Apr 2013 B2
8431766 Lonero Apr 2013 B1
8440877 Collins et al. May 2013 B2
8452388 Feldkamp et al. May 2013 B2
8471715 Solazzo et al. Jun 2013 B2
8507746 Ong et al. Aug 2013 B2
8518009 Saito Aug 2013 B2
8518010 Kuwano Aug 2013 B2
8546639 Wada et al. Oct 2013 B2
8563801 Berland et al. Oct 2013 B2
8570175 Rahimi Oct 2013 B2
8579876 Popp et al. Nov 2013 B2
8604268 Cohen et al. Dec 2013 B2
8623292 Song et al. Jan 2014 B2
8628506 Alex et al. Jan 2014 B2
8882731 Suzuki et al. Jan 2014 B2
8642832 Ales et al. Feb 2014 B2
8697933 Alex et al. Apr 2014 B2
8697934 Nhan et al. Apr 2014 B2
8697935 Daanen Apr 2014 B2
8698641 Abrham et al. Apr 2014 B2
8742198 Wei et al. Jun 2014 B2
8747379 Fletcher et al. Jun 2014 B2
8773117 Feldkamp et al. Jul 2014 B2
8779785 Wada et al. Jul 2014 B2
8785716 Schäfer et al. Jul 2014 B2
8816149 Richardson et al. Aug 2014 B2
8866052 Nhan et al. Oct 2014 B2
8866624 Ales et al. Oct 2014 B2
8884769 Novak Nov 2014 B2
8889944 Abraham et al. Nov 2014 B2
8920731 Nhan et al. Dec 2014 B2
8933291 Wei et al. Jan 2015 B2
8933292 Abraham et al. Jan 2015 B2
8962909 Groosman et al. Feb 2015 B2
8975465 Hong et al. Mar 2015 B2
8978452 Johnson et al. Mar 2015 B2
8988231 Chen Mar 2015 B2
9018434 Ruman Apr 2015 B2
9018435 Kawashima Apr 2015 B2
9034593 Martin et al. May 2015 B2
9070060 Forster Jun 2015 B2
9072632 Lavon Jul 2015 B2
9072634 Hundorf et al. Jul 2015 B2
9168185 Berland et al. Oct 2015 B2
9211218 Rinnert et al. Dec 2015 B2
9295593 Van Malderen Mar 2016 B2
9301884 Shah et al. Apr 2016 B2
9314381 Curran et al. Apr 2016 B2
9317913 Carney Apr 2016 B2
9380977 Abir Jul 2016 B2
9402771 Carney et al. Aug 2016 B2
9421137 LaVon et al. Aug 2016 B2
9545342 Cretu-petra Jan 2017 B2
9585795 Boase et al. Mar 2017 B2
10702705 Malchano Jul 2020 B2
20020021220 Dreyer Feb 2002 A1
20020070864 Jeutter et al. Jun 2002 A1
20030105190 Diehl et al. Jun 2003 A1
20030148684 Cramer et al. Aug 2003 A1
20030208133 Mault Nov 2003 A1
20040064114 David Apr 2004 A1
20040106202 Zainiev et al. Jun 2004 A1
20040127867 Odorzynski et al. Jul 2004 A1
20040127878 Olson Jul 2004 A1
20040220538 Panopoulos Nov 2004 A1
20040236302 Wilhelm et al. Nov 2004 A1
20040254549 Olson et al. Dec 2004 A1
20050008839 Cramer et al. Jan 2005 A1
20050033250 Collette Feb 2005 A1
20050065487 Graef et al. Mar 2005 A1
20050099294 Bogner May 2005 A1
20050107763 Matsuda May 2005 A1
20050124947 Femfors Jun 2005 A1
20050137542 Underhill et al. Jun 2005 A1
20050156744 Pires Jul 2005 A1
20050195085 Cretu-Petra Sep 2005 A1
20060058745 Pires Mar 2006 A1
20060061477 Yeh Mar 2006 A1
20060069362 Odorzynski Mar 2006 A1
20060195068 Lawando Aug 2006 A1
20060222675 Sahnis et al. Oct 2006 A1
20060224135 Beck Oct 2006 A1
20060229578 Roe Oct 2006 A1
20060264861 Lavon Nov 2006 A1
20070044805 Wedler Mar 2007 A1
20070046482 Chan Mar 2007 A1
20070055210 Kao Mar 2007 A1
20070100666 Stivoric May 2007 A1
20070142797 Long et al. Jun 2007 A1
20070156106 Klofta Jul 2007 A1
20070185467 Klofta et al. Aug 2007 A1
20070233027 Roe et al. Oct 2007 A1
20070252710 Long Nov 2007 A1
20070252711 Long et al. Nov 2007 A1
20070252713 Rondoni et al. Nov 2007 A1
20070255241 Weber et al. Nov 2007 A1
20070255242 Ales et al. Nov 2007 A1
20070282286 Collins Dec 2007 A1
20070287975 Fujimoto Dec 2007 A1
20080021423 Klofta Jan 2008 A1
20080021428 Klofta et al. Jan 2008 A1
20080052030 Olson et al. Feb 2008 A1
20080054408 Tippey et al. Mar 2008 A1
20080057693 Tippey et al. Mar 2008 A1
20080058740 Sullivan et al. Mar 2008 A1
20080058741 Long et al. Mar 2008 A1
20080058742 Ales Mar 2008 A1
20080074274 Hu Mar 2008 A1
20080082063 Ales Apr 2008 A1
20080132859 Pires Jun 2008 A1
20080147031 Long et al. Jun 2008 A1
20080208155 Lavon Aug 2008 A1
20080218334 Pitchers Sep 2008 A1
20080234644 Hansson et al. Sep 2008 A1
20080266117 Song et al. Oct 2008 A1
20080266122 Ales et al. Oct 2008 A1
20080266123 Ales Oct 2008 A1
20080269702 Ales Oct 2008 A1
20080269707 Song Oct 2008 A1
20080300559 Gustafson Dec 2008 A1
20080312622 Hundorf et al. Dec 2008 A1
20090058072 Weber et al. Mar 2009 A1
20090062756 Long et al. Mar 2009 A1
20090124990 Feldkamp et al. May 2009 A1
20090155753 Ales et al. Jun 2009 A1
20090198202 Nedestam Aug 2009 A1
20090275908 Song Nov 2009 A1
20090326409 Cohen et al. Dec 2009 A1
20090326504 Kaneda Dec 2009 A1
20100013778 Liu Jan 2010 A1
20100030173 Song et al. Feb 2010 A1
20100125949 Stebbirig May 2010 A1
20100145294 Song et al. Jun 2010 A1
20100152688 Handwerker et al. Jun 2010 A1
20100159599 Song et al. Jun 2010 A1
20100159611 Song et al. Jun 2010 A1
20100160882 Lowe Jun 2010 A1
20100164733 Ales et al. Jul 2010 A1
20100168694 Gakhar et al. Jul 2010 A1
20100168702 Ales et al. Jul 2010 A1
20100241094 Sherron Sep 2010 A1
20100277324 Yeh Nov 2010 A1
20110004175 Veith Jan 2011 A1
20110251038 Lavon Oct 2011 A1
20110298597 Kaihori Dec 2011 A1
20120310191 LaVon et al. Feb 2012 A1
20120061016 LaVon Mar 2012 A1
20120109087 Abraham May 2012 A1
20120116337 Ales May 2012 A1
20120116343 Yoshioka May 2012 A1
20120130330 Wilson et al. May 2012 A1
20120157947 Nhan Jun 2012 A1
20120161960 Cheng Jun 2012 A1
20120172824 Khaknazarov Jul 2012 A1
20120190956 Connolly Jul 2012 A1
20120206265 Solazzo Aug 2012 A1
20120225200 Mandzsu Sep 2012 A1
20120245541 Suzuki Sep 2012 A1
20120245542 Suzuki et al. Sep 2012 A1
20120256750 Novak Oct 2012 A1
20120282681 Teixeira et al. Nov 2012 A1
20120299721 Jones Nov 2012 A1
20120310190 LaVon et al. Dec 2012 A1
20120310192 Suzuki et al. Dec 2012 A1
20120323194 Suzuki et al. Dec 2012 A1
20130012896 Suzuki et al. Jan 2013 A1
20130018340 Abraham et al. Jan 2013 A1
20130023786 Mani et al. Jan 2013 A1
20130041334 Prioleau Feb 2013 A1
20130076509 Ahn Mar 2013 A1
20130110061 Abraham et al. May 2013 A1
20130110063 Abraham May 2013 A1
20130110075 Mukai May 2013 A1
20130131618 Abraham et al. May 2013 A1
20130151186 Feldkamp Jun 2013 A1
20130161380 Joyce et al. Jun 2013 A1
20130162402 Amann et al. Jun 2013 A1
20130162403 Stiemer et al. Jun 2013 A1
20130162404 Stiemer et al. Jun 2013 A1
20130165809 Abir Jun 2013 A1
20130211363 LaVon et al. Aug 2013 A1
20130261409 Pathak Oct 2013 A1
20130303867 Elfstrom et al. Nov 2013 A1
20130307570 Bosae et al. Nov 2013 A1
20130310796 Zink Nov 2013 A1
20130321007 Elfstrom et al. Dec 2013 A1
20130324955 Wong et al. Dec 2013 A1
20130338623 Kinoshita Dec 2013 A1
20140005020 LaVon et al. Jan 2014 A1
20140005622 Wirtz et al. Jan 2014 A1
20140014716 Joyce et al. Jan 2014 A1
20140015644 Amann et al. Jan 2014 A1
20140015645 Stiemer et al. Jan 2014 A1
20140022058 Stiemer et al. Jan 2014 A1
20140062663 Bourilkov et al. Mar 2014 A1
20140121473 Banet May 2014 A1
20140121487 Faybishenko et al. May 2014 A1
20140152442 Li Jun 2014 A1
20140155850 Shah et al. Jun 2014 A1
20140155851 Ales et al. Jun 2014 A1
20140163502 Arizti et al. Jun 2014 A1
20140188063 Nhan et al. Jul 2014 A1
20140198203 Vardi Jul 2014 A1
20140200538 Euliano et al. Jul 2014 A1
20140241954 Phillips et al. Aug 2014 A1
20140242613 Takeuchi et al. Aug 2014 A1
20140242715 Nhan et al. Aug 2014 A1
20140244644 Maschinchi et al. Aug 2014 A1
20140266736 Cretu-petra Sep 2014 A1
20140292520 Carney et al. Oct 2014 A1
20140306814 Ricci Oct 2014 A1
20140033442 Carney Nov 2014 A1
20140329212 Ruman et al. Nov 2014 A1
20140329213 Ruman et al. Nov 2014 A1
20140363354 Phillips et al. Dec 2014 A1
20140371702 Bosae et al. Dec 2014 A1
20150025347 Song Jan 2015 A1
20150042489 LaVon Feb 2015 A1
20150045608 Karp Feb 2015 A1
20150112202 Abir Apr 2015 A1
20150130637 Sengstaken, Jr. May 2015 A1
20150143881 Raut et al. May 2015 A1
20150150732 Abir Jun 2015 A1
20150157512 Abir Jun 2015 A1
20150206151 Carney et al. Jul 2015 A1
20150209193 Ying et al. Jul 2015 A1
20150223755 Abir Aug 2015 A1
20150276656 Striemer Oct 2015 A1
20150317684 Abir Nov 2015 A1
20150359490 Massey Dec 2015 A1
20160008182 Prokopuk et al. Jan 2016 A1
20160051416 Vartiainen et al. Feb 2016 A1
20160051417 Chu Feb 2016 A1
20160067113 Vartiainen et al. Mar 2016 A1
20160078716 Olafsson-Ranta et al. Mar 2016 A1
20160080841 Bergstrom et al. Mar 2016 A1
20160113822 Vartiainen et al. Apr 2016 A1
20160134497 Hermansson et al. May 2016 A1
20160136014 Arora et al. May 2016 A1
20160170776 Bergstrom et al. Jun 2016 A1
20160235603 Ehmsperger et al. Aug 2016 A1
20160287073 Pradeep Oct 2016 A1
20160292986 Pradeep Oct 2016 A1
20160293042 Pradeep Oct 2016 A1
20160345914 Jain Dec 2016 A1
20160374868 Ettrup Hansen Dec 2016 A1
20170035622 Wang Feb 2017 A1
20170108236 Guan Apr 2017 A1
20170116484 Johnson Apr 2017 A1
20170156594 Stivoric Jun 2017 A1
20170224543 Lavon Aug 2017 A1
20170224551 Lavon Aug 2017 A1
20170252225 Arizti Sep 2017 A1
20170278373 Ansley Sep 2017 A1
20170286977 Allen Oct 2017 A1
20180053396 Greene Feb 2018 A1
20180096290 Awad Apr 2018 A1
20180104114 Pepin Apr 2018 A1
20180149635 Abir May 2018 A1
20180204256 Bifolco Jul 2018 A1
20190180341 Matra Jun 2019 A1
20190336353 Arizti Nov 2019 A1
20200046574 Arizti Feb 2020 A1
20200060885 Arizti Feb 2020 A1
20200060886 Arizti Feb 2020 A1
Foreign Referenced Citations (64)
Number Date Country
103780814 May 2014 CN
203734736 Jul 2014 CN
203950108 Nov 2014 CN
104349040 Feb 2015 CN
205123860 Mar 2016 CN
105704451 Jun 2016 CN
106210543 Dec 2016 CN
106725506 May 2017 CN
206478417 Sep 2017 CN
207008159 Feb 2018 CN
0149880 May 1984 EP
1216673 Oct 2005 EP
1542635 Apr 2012 EP
249899 Jul 2014 EP
2549099 Oct 2017 GB
09187431 Jul 1997 JP
2002022687 Jan 2002 JP
2002143199 May 2002 JP
2003190209 Jul 2003 JP
2004230135 Aug 2004 JP
2006296566 Nov 2006 JP
WO 95016746 Jun 1995 WO
WO 99034841 Jul 1999 WO
0197466 Dec 2001 WO
2005011491 Feb 2005 WO
WO 2010123364 Oct 2010 WO
WO 2010123425 Oct 2010 WO
WO 2011013874 Feb 2011 WO
WO 2012084925 Jun 2012 WO
WO 2012126507 Sep 2012 WO
WO 2013003905 Jan 2013 WO
WO 2013016765 Feb 2013 WO
WO 2013061963 May 2013 WO
WO 2013091707 Jun 2013 WO
WO 2013091728 Jun 2013 WO
WO 2013095222 Jun 2013 WO
WO 2013095226 Jun 2013 WO
WO 2013095230 Jun 2013 WO
WO 2013095231 Jun 2013 WO
WO 2013097899 Jul 2013 WO
WO 2013181436 Dec 2013 WO
WO 2013185419 Dec 2013 WO
WO 2013189284 Dec 2013 WO
WO 2014035302 Mar 2014 WO
WO 2014035340 Mar 2014 WO
WO 2014122169 Aug 2014 WO
WO 2014137671 Sep 2014 WO
WO 2014146693 Sep 2014 WO
WO 2014146694 Sep 2014 WO
WO 2014148957 Sep 2014 WO
WO 2014177200 Nov 2014 WO
WO 2014177203 Nov 2014 WO
WO 2014177204 Nov 2014 WO
WO 2014177205 Nov 2014 WO
WO 2014178763 Nov 2014 WO
WO 2014192978 Dec 2014 WO
WO 2015003712 Jan 2015 WO
WO 2015068124 May 2015 WO
WO 2015102084 Jul 2015 WO
WO 2015102085 Jul 2015 WO
WO 2015127062 Aug 2015 WO
2016164373 Oct 2016 WO
2017217859 Dec 2017 WO
WO 2018216848 Nov 2018 WO
Non-Patent Literature Citations (11)
Entry
PCT International Search Report, dated Aug. 16, 2019, 14 pages.
All Office Actions, U.S. Appl. No. 16/402,348.
All Office Actions, U.S. Appl. No. 16/452,693.
All Office Actions, U.S. Appl. No. 16/452,712.
All Office Actions, U.S. Appl. No. 16/452,726.
Http://www.goodmonit.com/.
Https://techcrunch.com/2017/04/30/monit/ (May 1, 2017).
PCT Search Report, PCT/US2019/030302, dated Aug. 16, 2019, 14 pages.
All Office Actions, U.S. Appl. No. 17/220,356.
Unpublished U.S. Appl. No. 17/220,356, filed Apr. 1, 2021, to Blanca Arizti et al.
U.S. Appl. No. 16/402,348, filed May 3, 2019, Arizti, et al.
Related Publications (1)
Number Date Country
20200177848 A1 Jun 2020 US
Provisional Applications (1)
Number Date Country
62773401 Nov 2018 US