Remote monitoring

Information

  • Patent Grant
  • 8854194
  • Patent Number
    8,854,194
  • Date Filed
    Tuesday, March 26, 2013
    11 years ago
  • Date Issued
    Tuesday, October 7, 2014
    10 years ago
Abstract
Remote monitoring and inspection of measurement devices, emergency equipment, parking spaces, and other items is accomplished by using an image sensor (e.g., a CMOS sensor) to capture an image containing information about the monitored item. A signal containing information about the image (e.g., data representing the captured image or data indicating the state of the captured image) is transmitted to a remote central station.
Description
TECHNICAL FIELD

This disclosure relates to remote monitoring of measurement devices, emergency equipment, and other items.


BACKGROUND

Equipment, such as emergency equipment, fuel tanks, pipelines, etc., often includes one or more digital or analog measurement devices (e.g., liquid level indicators, pressure gauges, thermometers, flow meters, etc.) that provide important data about the equipment. It is also often important to ensure that equipment, such as fire extinguishers and other emergency equipment, are in its installed located and access to the equipment is not blocked. When a number of pieces of equipment are distributed across a large area or are located in an inaccessible areas, it becomes difficult to effectively monitor the status of the equipment.


SUMMARY

In general, in one aspect, an apparatus for detection of presence of an obstruction blocking access to or viewing of emergency equipment includes: an image sensor positioned to face an access area to the emergency equipment and configured to capture an image of the access area; an image recognition module configured to recognize at least whether the access area is obstructed and generate a first signal indicative of an obstructed state when the access area is recognized as obstructed; and communications circuitry configured to transmit to a remote central station a second signal including information indicative of the obstructed state of the emergency equipment provided by the captured image.


Embodiments may include one or more of the following features. The image sensor is configured to periodically capture an image of the access area. The access area includes an exit way. The image sensor is configured to capture an image of the access area in response to a change in a state of the emergency equipment. The change in a state of the emergency equipment comprises a removal of the emergency equipment from an installed position. The access area comprises a fire alarm pull station. Change in a state of the emergency equipment comprises an activation of the fire alarm pull station. The access area further comprises a portable defibrillator. The apparatus also includes a network linking the communication circuitry and the remote central station. The network comprises a wireless mesh network. The image sensor includes a CMOS sensor. The apparatus also includes a second image sensor positioned to face a display of the emergency equipment and configured to capture a second image of the display of the emergency equipment, wherein the communications circuitry is further configured to transmit to a remote central station a third signal including information about the display. The image recognition module is configured to recognize one of a plurality of predetermined states of the emergency equipment from the second image. The image recognition module is configured to recognized one of the plurality of predetermined states prior to transmission of the third signal to the remote central station. The emergency equipment includes a pressurized tank, and the display includes a pressure gauge on pressurized tank. The second image sensor is configured to periodically capture an image of the display. The apparatus also includes a second image sensor positioned to face an installed location of the emergency equipment and configured to capture a second image of the installed location, wherein the communications circuitry is further configured to transmit to the remote central station a third signal including information about the second image. The communication circuitry is configured to transmit a wireless signal to the remote central station. The network comprises a wireless mesh network. The network comprises an IEEE .1.-compliant network.


Other embodiments involve the foregoing features embodied as other apparatuses, methods, systems, software, hardware, or other computer products.


The details of one or more embodiments of the invention are set forth in the accompanying drawings and the description below. Other features, objects, and advantages of the invention will be apparent from the description and drawings, and from the claims.





DESCRIPTION OF DRAWINGS


FIGS. 1A-1C are diagrams of a digital image sensory device mounted on a pressure gauge of a fire extinguisher.



FIG. 2 is a diagram of a digital image sensory device in communication with a remote central station using a network.



FIGS. 3 and 5 are flow charts illustrating operation of digital image sensory devices.



FIG. 4 is a flow chart illustrated operation of a remote central station.



FIG. 6 is a diagram of multiple digital image sensory devices in communication with a central station using a mesh network.



FIGS. 7-9 and 11-12 are diagrams of digital image sensory devices configured to monitor various emergency equipment.



FIG. 10 is a diagram of a digital image sensory device configured to monitor parking spaces in a parking facility.





DETAILED DESCRIPTION

It is often desirable to remotely monitor digital or analog measurement devices, such as pressure gauges, liquid level indicators, thermometers, voltmeters, flow meters, etc., to ensure proper operation or check the status of important equipment (e.g., a fire extinguisher, portable oxygen tank, pipeline, battery, etc.), or measure usage of a consumable material (e.g., electric, gas, or water meter, liquid level indicator for a home heating oil or liquid propane tank, etc.). Often equipment will include a measurement device that can be viewed on site (e.g., a conventional pressure gauge on a fire extinguisher), but is incapable of transmitting a signal indicating the condition of the measurement device to a remote monitoring device. One way to remote monitoring of such a measurement device is to mount a digital image sensory device, such as a CMOS sensor, and communications circuitry on or near the measurement device. The digital image sensory device can be configured to capture an image of the measurement device at a desired frequency (e.g., every 1 second, 1 minute, 1 hour, 1 day, etc.) and can transmit the captured image to a remote monitoring station, either via wired or wireless transmission. Alternatively, the digital image sensor can include a microprocessor programmed to recognize predetermined conditions of a measurement device and transmit a signal indicating the presence of recognized predetermined conditions to a central station. Similarly, a digital image sensor can be configured to recognize the characters or numbers on the measurement device (e.g., a measurement device having a digital readout), decode them, and transmit a signal containing the decoded values.


For example, as shown in FIG. 1A, a fire extinguisher 10 includes a CMOS sensory device 12 attached to a pressure gauge 14, which monitors pressure of fire suppressant stored within the extinguisher's tank 16. The CMOS sensory device is attached in such a way that the CMOS camera (not shown) has a clear view to the display of the pressure gauge, but does not obstruct a person's view of the pressure gauge.


Referring to FIGS. 1B-1C, the CMOS sensory device 12 is releasably attached to the pressure gauge with a pair of suction cups 13a, 13b. In other implementations, the CMOS sensory device may be integral with the pressure gauge or attached to the pressure gauge with straps, clips, adhesive, screws, bolts, or other known releasable or non-releasable attachment devices.


The CMOS sensory device includes a CMOS camera 18 that is positioned to view a portion of the pressure gauge display that shows whether the indicator needle 19 is within a predetermined pressure range (shaded region R).


Referring to FIG. 2, the CMOS sensory device 12 includes a CMOS camera 18, microprocessor 21, communications circuitry and antenna 22, and battery 23 all mounted to a circuit board 24. The CMOS camera 18 is positioned to capture images of area A on the pressure gauge, which displays whether the indicator needle 19 is within the desired pressure range (e.g., shaded region R) or outside the desired pressure range. Battery 23 supplies power to the CMOS camera and other components on circuit board 24. Communications circuitry 22 provides two-way communication between the CMOS sensory device 12 and a remote central station 25 using a network 26 (e.g., a wireless IEEE 802.15.4-compliant or IEEE 802.11-compliant wireless network, Ethernet network, etc.). Microprocessor 21 controls operation of the CMOS sensory device by triggering the CMOS camera to periodically capture a digital image of area A on the pressure gauge. In some implementations, microprocessor 21 can be configured to “recognize” a state of the digital image and transmit a signal to the remote central station indicating the recognized state. In the illustrated example, the fire extinguisher's pressure gauge has at least two important states: (i) a normal state when the pressure gauge indicates a pressure within the operating limits and a (ii) an out-of-range state when the pressure gauge indicates a pressure outside the operating limits. Other implementations may recognize other states, such as a normal state, an above-normal state and a below-normal state. In some implementations, a digital image sensory device is configured to recognize alphanumeric characters displayed by a measurement device. For example, if a measurement device has a digital readout, a digital image sensory device may execute optical character recognition (OCR) software to recognize the alpha-numeric characters captured in the digital image.


The components of CMOS sensory device 12 shown in FIG. 2 are logical components, and, in actual implementations, one logical component may be implemented in separate physical component. Similarly, the functionality of multiple logical components may be combined and implemented as a single physical component. For example, a digital image sensory device may be actually implemented with multiple processors, e.g., one processor dedicated for the operation of the digital image sensor and a second processor dedicated to recognizing states and/or alpha-numeric characters captured in a digital image. Similarly, the functionality of communication circuitry and a microprocessor may be combined into a single physical component.


Referring to FIG. 3, a digital image sensory device, such as CMOS sensory device 12 shown in FIGS. 1-2, receives as input a time interval T (50) between image captures. In some implementations, the time interval may be preprogrammed and not variable by a user, while in other implementations the time interval may be programmed by a user using, for example, a program code contained in a separate microprocessor or on a microprocessor associated with wireless transmission circuitry.


Operation begins by starting (52) a timer and waiting until the timer reaches time T. When the timer reaches time T, the CMOS sensory device uses a CMOS camera to capture (54) an image of the measurement device. In addition, the timer is reset (56) and re-started (52) for the next image capture.


After the CMOS camera has captured an image of the measurement device, the CMOS sensory device analyzes (56) the captured image to determine (58) a state of the measurement device. In this example, the CMOS sensory devices analyzes the captured image to determine which of two states the image presents: (i) a “normal” state in which the image shows the measurement device to be within a predetermined pressure range, or (ii) an “out-of-range” state in which the image shown the measurement device to be outside a predetermined pressure range. The CMOS sensory device can use any known image recognition software, such as optical character recognition (OCR) or machine vision software, for determining the state of the measurement device shown in the image. If the CMOS sensory device cannot determine with a sufficient level of confidence (e.g., 90%) the state of the measurement device shown in the image, the CMOS sensory device reverts to an “error” state. In some implementations, a digital image sensory device may be programmed to capture and attempt to recognize another digital image, or, alternative, transmit the unrecognizable image back to the remote central station, if the digital image sensory device reverts to an error state.


If the CMOS sensory device determines from the digital image that the measurement device is within the predetermined pressure range (i.e., in the normal state), it transmits (60) a signal to a remote central station indicating that the measurement device is within the predetermined pressure range along with a time stamp of when the digital image was captured and an identification code associated with the CMOS sensory device. If the CMOS sensory device determines from the digital image that the measurement device is out of the predetermined pressure range (i.e., in the out-of-range state), then it transmits (62) a signal to the remote central station indicating that the measurement device is out of the predetermined pressure range, along with a time stamp of when the digital image was captured and the CMOS sensory device's identification code. Similarly, if the CMOS sensory device cannot determine from the digital image whether the measurement device is within or out of the predetermined pressure range (i.e., in the error state), transmits (64) a signal to the remote central station indicating an error, along with a time stamp of when the digital image was captured and the CMOS sensory device's identification code.


When a signal is received at the remote central station, the central station decodes and stores the transmitted information and automatically notifies appropriate personnel (e.g., fire department, police department, building management department, etc.) of abnormal conditions. For example, as shown in FIG. 4, a remote central station receives (68) a signal from one of several CMOS sensory devices distributed throughout a room, building, group of buildings, or other geographic area. The signal preferably includes information about the state of the monitored measurement device determined from a captured digital image of the device, the time at which the digital image was captured, and a unique identification code of the CMOS sensory device. The remote central station decodes and stores (70) the received sensory data in a database, data table or similar data structure. The remote central station also accesses a registry (74) of CMOS sensory devices, which matches each CMOS sensory device by its unique identification code. The remote central station accesses the registry of CMOS sensory device to associate (72) the received signal with its corresponding measurement device. For example, a CMOS sensory device having unique ID code “123” may be attached to a pressure gauge on a fire extinguisher located on the second floor, room 205 of a particular building. The CMOS registry includes, for example, a table that associates the unique ID code “123” with the type (fire extinguisher pressure gauge) and location (second floor, room 205) of the corresponding measurement device.


The remote central station also displays the received data on a display device. In some implementations, the received data is overlaid on a graphical map of the building, floor, campus, etc. that is being monitored. To allow a user to more quickly determine the status of an array of distributed measurement devices monitored by CMOS sensory devices, the remote central station may display color codes (e.g., red for a current abnormal reading, green for a current normal reading, and yellow for a current error or warning reading) on the graphical map at the locations where the measurement devices are located or installed. In some implementations, the central station employs a glanceable display device, such as an Ambient Orb™ or Ambient Dashboard™ by Ambient Devices (www.ambientdevices.com). A glanceable device is a device, such as a light source, meter, or sound source, which communicates information without demanding a user's attention. For example, status information on monitored measurement devices could be processed for output to a glanceable display such that monitored devices at a location (e.g., in a room, on a floor, in a building, etc.) or of a particular type (e.g., fire extinguisher pressure gauges, portable oxygen tank pressure gauges, etc.) causes a light to glow more green in color as the more devices show a normal status and glow more red in color as they show a non-normal status (e.g., abnormal or error). Similarly, sensory data indicating a normal status of monitored measurement devices may cause a needle to move towards one end of a scale whereas as monitored devices show a abnormal status the needles moves towards an opposite end of the scale. In this way, sensory data collected from many measurement devices can be analyzed in the aggregate in a simple, easy-to-view format.


It may be desirable for a user at a remote central station to download captured images for inspection. For example, if a CMOS sensory device transmits a signal indicating an out-of-range pressure condition or an error condition, an operator at the remote central station may want to view the captured image to verify the reading or to see if the human operator can read the captured image. Similarly, an operator may want to periodically test operation of the system by comparing captured images with the state indicated by the CMOS sensory device. Accordingly, in some implementations, a CMOS sensory device also includes on-board memory for storing one or more captured digital images and is adapted to respond to requests from a remote central station to upload one or more of the images stored in the memory to the central station. As the CMOS sensory device captures a new image, it overwrites the oldest image stored in its memory. If a user at a remote central station wants to view one or more of the images stored at the CMOS sensory device, the user is able to send a command to the CMOS sensory device instructing it to send one or more of the stored images. In some implementations, the CMOS sensory device may also be equipped to receive a command from a remote user instructing the CMOS sensory device to immediately capture and analyze a monitored measurement device. Thus, a remote user is able to check on the status of a monitored measurement device without having to wait for the next time at which the CMOS sensory device is programmed to capture a digital image.


In some implementations, the CMOS sensory device does not include any image recognition software for determining the state of the measurement device as shown in a captured digital image. For example, as shown in FIG. 5, a CMOS sensory device is configured to capture a digital image of a measurement device every T time interval (90-96), and, then transmits (98) the captured digital image directly to the remote central station along with a time stamp and an identification code associated with the CMOS sensory device. Image recognition software may be employed at the remote central station to analyze the incoming images to determine a state of a measurement device shown in the image. Alternatively, the remote central station may not include any image recognition software, but may simply store and/or display captured images for review by a user.


Referring to FIG. 6, a wireless mesh network 100, such as a Zigbee™ wireless mesh network or other IEEE 802.15.4-compliant wireless mesh network, is used to transmit information between a remote central station 102 and remote CMOS sensory devices (not shown) mounted to portable oxygen tanks 104, 106 used, for example, at a hospital, nursing home, or other healthcare facility. The wireless mesh network includes several nodes 108-119, which exchange wireless signals between CMOS sensory devices located within the mesh network and a remote central station 102. In some implementations, the remote central station may include one or more devices for transmitting sensory data to another device (e.g., a server computer) located outside the wireless mesh network. For example, the remote central station may be equipped with a cellular modem, satellite modem, public switched telephone network (PSTN) modem, and/or networking card transmitting data to a remote server. The remote central station may also be equipped to receive and execute commands from a remote device (e.g., a command to have one or more CMOS sensory devices capture and analyze a digital image of a measurement device. The remote central station may be implemented, for example, as a laptop computer, desktop computer, or a single-board computer (e.g., cellular telephone, PDA, etc.).


In some implementations, other types of communication networks than a wireless mesh network are used to exchange. For example, a Bluetooth-compliant wireless network or other IEEE 802.11-compliant wireless network or a hardwire network, such as an Ethernet network, may be used to exchange information between CMOS sensory device and a remote central station.


In addition to using digital image sensory devices to remotely monitor measurement devices (e.g., pressure gauges, fuel level indicators, thermometers, flow meters, etc.), digital image sensors can be used to detect presence of important equipment, such as a portable fire extinguisher or portable defibrillator, at a predetermined location. In addition, digital image sensors can be used to detect presence of an obstruction blocking access to important equipment, such as a fire extinguisher, emergency exit, fire alarm pull station, defibrillator, etc.


For example, as shown in FIG. 7, a fire extinguisher 120 is mounted to a wall, post, or other support surface, W. A CMOS sensory device 122 is mounted on the wall behind the fire extinguisher 120 with two screws 121a, 121b. The CMOS sensory device includes a CMOS camera 122 and houses components shown in FIG. 2, including a microprocessor, communications circuit, and battery mounted to a circuit board. The CMOS camera is positioned to face the fire extinguisher, and, is configured to periodically capture a digital image of the fire extinguisher at its installed position. By periodically capturing digital images of the fire extinguisher in its installed position, the CMOS sensory device can detect presence (or lack thereof) of the fire extinguisher at its installed location.


In some implementations, the CMOS sensory device includes optical recognition software that recognizes a state of a captured digital image (e.g., installed position state, missing from position state) and transmits a signal indicating the determined state to a remote central station. As described above, the CMOS sensory device in some implementations also includes on-board memory for storing one or more captured digital images and is adapted to respond to requests from a remote central station to upload one or more of the images stored in the on-board memory to the central station. Similarly, in some implementations, the CMOS sensory device may also be equipped to receive a command from a remote user instructing the CMOS sensory device to immediately capture and analyze a monitored measurement device.


Rather than transmitting a status of a captured digital image, the CMOS sensory device may be configured to transmit the digital image to the remote central station, where it may be displayed for a human to analyze, analyzed by image recognition software, or send to another location (e.g., a network server accessible over a public or private network) for display or analysis.


As mentioned above, a digital image sensory device may also be used to detect obstructions to viewing of or obstruction to important equipment, such as emergency equipment. For example, as shown in FIG. 8, a cabinet 132 housing a fire extinguisher 130 includes two CMOS sensors 134, 136. The CMOS camera of CMOS sensor 134 faces the installed position of the fire extinguisher and, therefore, detects when the fire extinguisher is removed from its installed position. The CMOS camera of CMOS sensor 136 faces outward and periodically captures digital images of an access area immediately in front of the cabinet to detect when an object is blocking access to the cabinet. Similarly, as shown in FIG. 9, an emergency exit door includes several CMOS sensors 140a-140c with CMOS cameras that periodically capture images of areas in front of the door to detect presence of obstructions 142, 144 to the emergency exit door. Addition CMOS sensory devices may be placed on the other side of the door 138 to detect presence of obstructions on the opposite side of the door. The CMOS sensory device may be configured to transmit captured digital images to a remote central station via a wireless or hardwired network or may include image recognition software to determine a state of the captured digital image (e.g., no obstruction state, obstructed state, reading error) and transmit a signal indicated the detected state to the remote central station.


CMOS sensory devices can be used presence of any item in a predetermined position such as a portable defibrillator, life jackets/preservers, wheeled fire extinguisher, or an automobile in a parking space. Similarly, CMOS sensory devices can be used to detect obstruction to any item or object in which viewing and/or physical access is important (e.g., fire alarm pull stations, emergency lighting, defibrillators, emergency stairways and exitways, fire hose reels, fire enunciator panels, loading docks, or fire lanes.


As shown in FIG. 10, a system 145 for monitoring status of vehicles 146a-146f parked in a parking facility includes a several digital image sensory devices 147a-147h that each face a parking space 148a-148h. The digital sensory devices are in communication with a remote central station 149 over a wireless network (not shown) such as a Zigbee™ wireless mesh network. Each digital image sensory device is programmed to periodically (e.g., every 5 seconds) to capture a digital image of the parking space, determine the state of the space (e.g., empty or occupied), and transmit to a central station a signal identifying the digital image sensory device and the determined state of the space. At the central station 149, signals from each of the digital sensory devices are received and assimilated to determine capacity of the parking facility. In the illustrated example, a worker could view a display on the remote central station and direct an entering customer to the location of one of the two open spaces (i.e., space 148c and 148e) in the facility. In some implementations, the digital image sensory devices may transmit the digital image to the central station where it is analyzed (either by software or by human) to determine the state of the parking space.


In some implementations, a digital image sensory device is used to capture images of the area around a piece of portable equipment when the equipment is moved from its installed position to identify a person who removed the equipment, and, deter theft or improper use of the equipment. For example, as shown in FIG. 11, a fire alarm pull station 150 is mounted to a wall W or other structure. The fire alarm pull station 150 includes a build-in CMOS sensory device 152, in which a CMOS camera faces an area in front of the pull station. The CMOS sensory device is configured to capture one or more digital images and store when the fire alarm pull station is activated (i.e., the handle 154 is pulled down). The CMOS sensory device either stores the captured digital image for later retrieval or automatically transmits the captured image to the remote central station. Because the intent of the captured digital images immediately following activation of the pull station is to determine the identify of the person who pulled the handle, the CMOS sensory device need not apply image recognition software to determine a state of the image. However, in some implementations, the CMOS sensory device may periodically capture digital images of the area in front of the pull station in order to detect presence of an obstruction. For these digital images, the CMOS sensory device may apply image recognition software to determine a state of the pull station (e.g., pull station obstructed, pull station not obstructed, recognition error).


Similarly, as shown in FIG. 12, a fire extinguisher 156 is installed in a cabinet 158 mounted within a wall W. A CMOS sensory device 160 is built into a housing 162. The CMOS sensory device includes a CMOS camera positioned to face an area outside the cabinet at between 4 and 7 feet off the ground (the approximate height of a person's head). A tether 164 is attached between the fire extinguisher 156 and the housing 162 to detect removal of the fire extinguisher from its installed position. U.S. Pat. Nos. 5,484,651, 6,302,218, 6,585,055 and 6,488,099, each of which are incorporated herein by reference, describes examples of tethers used to detect removal of equipment from an installed position. Upon detection of removal of fire extinguisher 156 from its installed position, the CMOS sensory device 160 is configured to capture one or more digital images of the area in front of the cabinet. The CMOS sensory device can be configured to store the digital image(s) for later retrieval and/or automatically transmit captured images to a remote central station, where they may be analyzed to determine the identity of the person that removed the extinguisher.


A number of embodiments of the invention have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the invention. For example, while a CMOS camera has been illustrated and described as part of the digital image sensor, other digital image capturing devices such as CCD (Charge Coupled Device), full frame, interline, or linear image sensor may be used. Accordingly, other embodiments are within the scope of the following claims.

Claims
  • 1. Apparatus for detection of presence of an obstruction blocking access to an emergency exit, the apparatus comprising: an image sensor positioned to face an access area to the emergency exit and configured to capture an image of the access area;an image recognition module configured to: recognize at least whether the access area is obstructed; andgenerate a first signal indicative of an obstructed state when the access area is recognized as obstructed; andcommunications circuitry configured to transmit to a remote central station a second signal including information indicative of the obstructed state of the emergency exit provided by the captured image.
  • 2. The apparatus of claim 1 wherein the image sensor is configured to periodically capture an image of the access area.
  • 3. The apparatus of claim 1 wherein the emergency exit includes an emergency stairway.
  • 4. The apparatus of claim 1 wherein the image sensor is configured to capture an image of the access area in response to a change in a state of an item of emergency equipment.
  • 5. The apparatus of claim 4 wherein the change in a state of the item of emergency equipment comprises a removal of the emergency equipment from an installed position.
  • 6. The apparatus of claim 4 wherein the access area includes a fire alarm pull station.
  • 7. The apparatus of claim 6 wherein the change in the state of the emergency equipment comprises an activation of the fire alarm pull station.
  • 8. The apparatus of claim 1 wherein the access area includes a portable defibrillator.
  • 9. The apparatus of claim 1 further comprising a network linking the communication circuitry and the remote central station.
  • 10. The apparatus of claim 9 wherein the network includes a wireless mesh network.
  • 11. The apparatus of claim 1 wherein the image sensor includes a CMOS sensor.
  • 12. The apparatus of claim 1 further comprising a second image sensor positioned to face a display of an item of emergency equipment in the access area and configured to capture a second image of the display of the item of emergency equipment, wherein the communications circuitry is further configured to transmit to a remote central station a third signal including information about the display.
  • 13. The apparatus of claim 12 wherein the image recognition module is configured to recognize one of a plurality of predetermined states of the item of emergency equipment from the second image.
  • 14. The apparatus of claim 13 wherein the image recognition module is configured to recognized one of the plurality of predetermined states prior to transmission of the third signal to the remote central station.
  • 15. The apparatus of claim 12 wherein the item of emergency equipment includes a pressurized tank, and the display includes a pressure gauge on pressurized tank.
  • 16. The apparatus of claim 12 wherein the second image sensor is configured to periodically capture an image of the display.
  • 17. The apparatus of claim 1 further comprising a second image sensor positioned to face an installed location of an item of emergency equipment and configured to capture a second image of the installed location, wherein the communications circuitry is further configured to transmit to the remote central station a third signal including information about the second image.
  • 18. The apparatus of claim 1 wherein the communication circuitry is configured to transmit a wireless signal to the remote central station.
  • 19. The apparatus of claim 1 wherein the network includes a wireless mesh network.
  • 20. The apparatus of claim 16 wherein the network includes an IEEE 802.15.4-compliant network.
RELATED APPLICATION INFORMATION

This application is a continuation of U.S. application Ser. No. 13/451,819 filed Apr. 20, 2012, which is a continuation of U.S. application Ser. No. 13/196,371 filed Aug. 2, 2011, now U.S. Pat. No. 8,248,216 issued Aug. 21, 2012, which is a continuation of U.S. application Ser. No. 12/716,366 filed Mar. 3, 2010, now U.S. Pat. No. 8,009,020 issued Aug. 30, 2011, which is a continuation of U.S. application Ser. No. 11/071,132 filed Mar. 2, 2005, now U.S. Pat. No. 7,728,715 issued Jun. 1, 2010, each of which is incorporated herein by reference in its entirety.

US Referenced Citations (113)
Number Name Date Kind
922456 Casey May 1909 A
2670194 Hansson Feb 1954 A
3145375 Webb Aug 1964 A
3333641 Hansom Aug 1967 A
3664430 Sitabkhan May 1972 A
3735376 Kermer et al. May 1973 A
3946175 Sitabkhan Mar 1976 A
4003048 Weise Jan 1977 A
4015250 Fudge Mar 1977 A
4034697 Russell Jul 1977 A
4051467 Galvin Sep 1977 A
4100537 Carlson Jul 1978 A
4101887 Osborne Jul 1978 A
4119153 Avant Oct 1978 A
4125084 Salmonsen et al. Nov 1978 A
4143545 Sitabkhan Mar 1979 A
4184377 Hubbard Jan 1980 A
4246046 Lameyer Jan 1981 A
4279155 Balkanli Jul 1981 A
4289207 Wernert Sep 1981 A
4303395 Bower Dec 1981 A
4342988 Thompson et al. Aug 1982 A
4360802 Pinto Nov 1982 A
4418336 Taylor Nov 1983 A
4419658 Jarosz et al. Dec 1983 A
4531114 Topol et al. Jul 1985 A
4548274 Simpson Oct 1985 A
4586383 Blomquist May 1986 A
4599902 Gray Jul 1986 A
4613851 Hines Sep 1986 A
4697643 Sassier Oct 1987 A
4805448 Armell Feb 1989 A
4823116 Kitchen, III et al. Apr 1989 A
4823788 Smith et al. Apr 1989 A
4835522 Andrejasich et al. May 1989 A
4866423 Anderson et al. Sep 1989 A
4887291 Stillwell Dec 1989 A
4890677 Scofield Jan 1990 A
4928255 Brennecke May 1990 A
4979572 Mikulec Dec 1990 A
5123409 Sheffield et al. Jun 1992 A
5153567 Chimento Oct 1992 A
5153722 Goedeke et al. Oct 1992 A
5224051 Johnson Jun 1993 A
5357242 Morgano et al. Oct 1994 A
5388570 Wassil Feb 1995 A
5400246 Wilson et al. Mar 1995 A
5460228 Butler Oct 1995 A
5475614 Tofte et al. Dec 1995 A
5486811 Wehrle et al. Jan 1996 A
5534851 Russek Jul 1996 A
5578993 Sitabkhan et al. Nov 1996 A
5593426 Morgan et al. Jan 1997 A
5596501 Comer et al. Jan 1997 A
5613778 Lawson Mar 1997 A
5652393 Lawson Jul 1997 A
5706273 Guerreri Jan 1998 A
5775430 McSheffrey Jul 1998 A
5781108 Jacob et al. Jul 1998 A
5793280 Hincher Aug 1998 A
5808541 Golden Sep 1998 A
5848651 McSheffrey et al. Dec 1998 A
5853244 Hoff et al. Dec 1998 A
5864287 Evans, Jr. et al. Jan 1999 A
5877426 Hay et al. Mar 1999 A
5936531 Powers Aug 1999 A
5952919 Merrill Sep 1999 A
6014307 Crimmins Jan 2000 A
6114823 Doner et al. Sep 2000 A
6125940 Oram Oct 2000 A
6128576 Nishimoto et al. Oct 2000 A
6141584 Rockwell et al. Oct 2000 A
6155160 Hochbrueckner Dec 2000 A
6168563 Brown Jan 2001 B1
6240365 Bunn May 2001 B1
6270455 Brown Aug 2001 B1
6289331 Pedersen et al. Sep 2001 B1
6301501 Cronin et al. Oct 2001 B1
6302218 McSheffrey et al. Oct 2001 B1
6311779 McSheffrey et al. Nov 2001 B2
6317042 Engelhorn et al. Nov 2001 B1
6327497 Kirchgeorg et al. Dec 2001 B1
6336362 Duenas Jan 2002 B1
6351689 Carr et al. Feb 2002 B1
6357292 Schultz et al. Mar 2002 B1
6401713 Hill et al. Jun 2002 B1
6450254 Hoyle et al. Sep 2002 B1
6488099 McSheffrey et al. Dec 2002 B2
6496110 Peterson et al. Dec 2002 B2
6542076 Joao Apr 2003 B1
6585055 Mcsheffrey et al. Jul 2003 B2
6587049 Thacker Jul 2003 B1
6598454 Brazier et al. Jul 2003 B2
6646545 Bligh Nov 2003 B2
6735473 Kolder et al. May 2004 B2
6853309 Schröter Feb 2005 B1
6856251 Tietsworth et al. Feb 2005 B1
7271704 McSheffrey et al. Sep 2007 B2
7728715 Riedel et al. Jun 2010 B2
7891435 McSheffrey et al. Feb 2011 B2
8009020 Riedel et al. Aug 2011 B2
8248216 Riedel et al. Aug 2012 B2
20010052681 Deavila Dec 2001 A1
20030071736 Brazier et al. Apr 2003 A1
20030116329 McSheffrey et al. Jun 2003 A1
20030135324 Navab Jul 2003 A1
20030189492 Harvie Oct 2003 A1
20040017471 Suga et al. Jan 2004 A1
20100171624 McSheffrey et al. Jul 2010 A1
20100245570 Riedel et al. Sep 2010 A1
20110109454 McSheffrey, Sr. et al. May 2011 A1
20110285847 Riedel et al. Nov 2011 A1
20120200712 Riedel et al. Aug 2012 A1
Foreign Referenced Citations (10)
Number Date Country
3731793 Mar 1989 DE
2340109 Sep 1977 FR
2515845 May 1983 FR
2676931 Dec 1992 FR
WO-8102484 Sep 1981 WO
WO-9411853 May 1994 WO
WO-0146780 Jun 2001 WO
WO-0193220 Dec 2001 WO
WO-03076765 Sep 2003 WO
WO-03098908 Nov 2003 WO
Non-Patent Literature Citations (13)
Entry
“U.S. Appl. No. 12/716,366, Non-Final Office Action mailed Sep. 15, 2010”, 3 Pgs.
“U.S. Appl. No. 13/196,371), Notice of Allowance mailed May 23, 2012”, 7 pages.
“NFPA 10 Standard for Portable Fire Extinguishers”, National Fire Protection Association, 1988 edition 1998 , pp. 10.1-10.56.
“U.S. Appl. No. 10/614,948, Notice of Allowance mailed Dec. 8, 2010”, 7 pages.
“U.S. Appl. No. 11/856,618, Non-Final Office Action mailed Sep. 28, 2010”, 6 pgs.
“PCT/US2004/022019 International Search Report and the Written Opinion”, Dec. 6, 2004, 11 pages.
““Exciting new Products for Measuring Flow and Pressure””, Cole-Parmer Brochure Canada, Apr. 23, 1996 , 1 pg.
, ““Help That Comes Too Late is as Good as No Help at All—The Fire Extinguisher Alarm System Gives Immediate Help””, Invention Technologies, Inc. Press Release, Undated, 1 page.
“U.S. Appl. No. 12/716,366, Notice of Allowance mailed Apr. 29, 2011”, , 6 pages.
, “U.S. Appl. No. 13/196,371, Non-Final Office Action mailed Oct. 13, 2011”, 11 pages.
“U.S. Appl. No. 13/451,819 Notice of Allowance mailed Feb. 4, 2013”, 5 pages.
“U.S. Appl. No. 13/451,819, Non-Final Office Action mailed Sep. 14, 2012”, 17 pages.
“Application Serial No. 2,443,773, Office Action dated Dec. 10, 2008”, 2 pages.
Related Publications (1)
Number Date Country
20130286201 A1 Oct 2013 US
Continuations (4)
Number Date Country
Parent 13451819 Apr 2012 US
Child 13850401 US
Parent 13196371 Aug 2011 US
Child 13451819 US
Parent 12716366 Mar 2010 US
Child 13196371 US
Parent 11071132 Mar 2005 US
Child 12716366 US