Autonomous coverage robot navigation system

Information

  • Patent Grant
  • 8380350
  • Patent Number
    8,380,350
  • Date Filed
    Tuesday, December 23, 2008
    16 years ago
  • Date Issued
    Tuesday, February 19, 2013
    11 years ago
Abstract
An autonomous mobile robot system for bounded areas including a navigation beacon and an autonomous coverage robot. The navigation beacon has a gateway beacon emitter arranged to transmit a gateway marking emission with the navigation beacon disposed within a gateway between the first bounded area and an adjacent second bounded area. The autonomous coverage robot includes a beacon emission sensor responsive to the beacon emission, and a drive system configured to maneuver the robot about the first bounded area in a cleaning mode in which the robot is redirected in response to detecting the gateway marking emission. The drive system is also configured to maneuver the robot through the gateway into the second bounded area in a migration mode.
Description
TECHNICAL FIELD

This invention relates to robots, and more particularly to autonomous coverage robots and associated navigation systems.


BACKGROUND

Autonomous robots are robots which can perform desired tasks in unstructured environments without continuous human guidance. Many kinds of robots are autonomous to some degree. Different robots can be autonomous in different ways. An autonomous coverage robot traverses a work surface without continuous human guidance to perform one or more tasks. In the field of home, office and/or consumer-oriented robotics, mobile robots that perform household functions such as vacuum cleaning, floor washing, patrolling, lawn cutting and other such tasks have been widely adopted.


SUMMARY

In one aspect, an autonomous coverage robot includes a chassis defining a forward drive direction, a controller carried by the chassis, omni-directional receiver carried by the chassis, and a directional receiver disposed on a forward portion of the chassis and responsive to an emission incident on the forward portion of the chassis from the drive direction. The directional receiver includes a mounting receptacle defining first and second apertures substantially aligned with the drive direction, and first and second component receivers housed in the receptacle and positioned to be responsive to the emission received through the first and second apertures, respectively. The component receivers are each configured to generate a respective signal in response to the emission. The controller is configured to determine a direction of the emission in relation to the drive direction based on the signals generated by the component receivers.


The directional receiver is installed onto a front portion of the chassis and housed within the mounting receptacle. The mounting structure stabilizes and supports the directional receiver which is used for docking and navigation throughout a room. The directional receiver includes two detectors which are highly directional in the sense that they are able to detect peaks of an overlapping signal from far away and servo in on the source of the beam using those peaks. In one example, the first and second component receivers of the directional receiver are positioned to form an angle of between about 1 and 15 degrees there between. In addition, the first and second component receivers may be collimated to focus at a point between about 3 and 5 meters from the robot or on any point within each bounded area.


The controller maneuvers the robot to detect an emission with the omni-directional receiver, and in response to the detection, orients the chassis to align the robot drive direction with the determined emission direction based on the signals generated by the component receivers of the directional receiver. The controller, in response to a detected emission with the directional receiver, orients the chassis to align the robot drive direction with the determined emission direction based on the signals generated by the component receivers of the directional receiver. The receivers may be configured to receive transmissions of infrared light.


In one implementation, the omni-directional receiver includes a housing having an upper portion and defining an inner cavity, a conical, and an emission receiver. The upper portion allows a transmission of an emission into the inner cavity. The conical reflector is disposed on an upper surface of the cavity to reflect emissions incident on the upper portion of the housing down into the cavity. The emission receiver is disposed in the cavity below the conical reflector.


In one example, the robot includes a floor cleaning assembly carried by the chassis, a cleaning bin carried by the chassis and arranged to collect debris removed from a floor by the cleaning assembly, and a bin sensor. The bin sensor is configured to generate a signal indicative of the cleaning bin reaching a threshold filling that causes the controller to initiate a docking sequence.


In another aspect, an autonomous mobile robot system for bounded areas includes a navigation beacon and an autonomous coverage robot. The navigation beacon has a gateway beacon emitter arranged to transmit a gateway marking emission with the navigation beacon disposed within a gateway between the first bounded area and an adjacent second bounded area. The autonomous coverage robot includes a beacon emission sensor responsive to the beacon emission, and a drive system configured to maneuver the robot about the first bounded area in a cleaning mode in which the robot is redirected in response to detecting the gateway marking emission. The drive system is also configured to maneuver the robot through the gateway into the second bounded area in a migration mode.


The beacon may be configured to emit an infrared signal for the gateway marking emission, or any other type of signal, which is readily stopped by a wall so as not to “bleed over” into adjacent rooms and uniquely identify the current room or locality in which the robot is located.


In one implementation, the robot remains in the first bounded area upon the robot redirection in response to detecting the gateway marking emission in the cleaning mode. The gateway marking emission may be infrared light and the beacon emission sensor is configured to detect transmissions of infrared light.


In another implementation, the drive system is configured to maneuver the robot about the first bounded area in the cleaning mode for a preset time interval, and to automatically initiate the migration mode at expiration of the time interval. The drive system is configured to maneuver the robot across the gateway emission in the migration mode.


In yet another implementation, the drive system is configured to maneuver the robot about the first bounded area in the cleaning mode until a preset number of encounters of the gateway marking emission have been detected, and then to automatically initiate the migration mode. The coverage robot's dwell time in one room or selective rooms can also be set by a timer or schedule, a number of collisions or interactions with a proximity beam of a particular beacon, a character or number of dirt or debris detections, remaining battery life, and maintenance or remote control overrides.


In one example, the robot includes a floor cleaning system that removes debris from the floor as the robot is maneuvered.


In some examples, the robot includes a transmitter that signals the beacon to disrupt transmission of the gateway emission in the migration mode.


In another example, the navigation beacon includes a vectoring beacon emitter arranged to transmit a directed vectoring emission into the first bounded area with the beacon disposed within the gateway. The robot drive system is configured to direct the robot toward the beacon upon encountering the vectoring emission in the migration mode. The navigation beacon transmits the directed vectoring emission, which may be infrared light, at an angle adjacent the cross-gateway direction of between about 45 and 90 degrees. The robot, while in the migration mode, discerns a position of the navigation beacon in response to detection of the directed vectoring emission and aligns a drive direction defined by the robot in relation to an emission path defined by the directed vectoring emission. The robot advances along the emission path to reach and traverse the gateway. The navigation beacon may also transmit a proximity emission laterally about the beacon, where the robot avoids cleaning and migration within the proximity emission.


In some examples, the robot wirelessly communicates with the navigation beacon to deactivate the directed vectoring emission while in the cleaning mode and activate the directed vectoring emission while in the migration mode. In other examples, the robot remotely activates the directed vectoring emission of the navigation beacon and deactivates the gateway emission upon initiating the migration mode. Furthermore, the robot remotely deactivates the directed vectoring emission of the navigation beacon and activates the gateway emission upon terminating the migration mode. The robot remotely activates and deactivates the emissions of the navigation beacon with a radiofrequency communication. The navigation beacon may also be configured to communicate schedule information with the robot.


The system may also include a base station located in one of the bounded areas. The base station includes a base defining a docking direction from which the robot may properly dock, and a robot charger housed in the base. An omni-directional beam emitter may be mounted on the base and configured to project a proximity beam laterally around the docking station. Two navigational field emitters are housed in the base and arranged to emit respective, laterally bounded and overlapping fields of emissions of signal beams, respectively. One of the emitted fields defines a lateral field edge aligned with the docking direction and overlapped by the other of the fields. The two navigational field emitters of the base station are positioned to form an angle therebetween of between about 45 and 90 degrees. The fields of emissions of the two navigational field emitters of the base station may be infrared light.


In another aspect, a method of navigating an autonomous coverage robot between bounded areas includes positioning a navigation beacon in a gateway between adjoining first and second bounded areas. The beacon configured to transmit a gateway marking emission across the gateway. In some example, the navigation beacon may also transmit a proximity emission laterally about the beacon, where the robot avoids cleaning and migration within the proximity emission. The method also includes placing the coverage robot within the first bounded area. The robot autonomously traverses the first bounded area in a cleaning mode and upon encountering the gateway marking emission in the gateway, the robot remains in the first bounded area, thereby avoiding the robot migration into the second area. Upon termination of the cleaning mode in the first area, the robot autonomously initiates a migration mode to move through the gateway, past the beacon, into the second bounded area.


In some examples, the robot remotely deactivates the gateway marking emission of the navigation beacon upon initiating the migration mode and activates the gateway marking emission upon terminating the migration mode. The gateway marking emission may be infrared light. In other examples, the robot is non-responsive to the gateway emission while in the migration mode.


In one example, upon termination of the migration mode in the second area, the robot autonomously initiates the cleaning mode in the second bounded area.


In one example, the navigation beacon is configured to transmit a directed vectoring emission into the first bounded area with the beacon disposed within the gateway. The robot drives toward the beacon upon detecting the vectoring emission in the migration mode. Detecting the directed vectoring emission in the migration mode includes the robot sensing the directed vectoring emission with a directional receiver on the robot, where the directional receiver is aligned with a robot drive direction, and the robot aligning the robot drive direction with a path defined by the directed vectoring emission. The robot may also sense the directed vectoring emission with an omni-directional receiver on the robot before maneuvering to detect the directed vectoring emission by the directional receiver on the robot. The robot moves past the beacon by moving towards the beacon along the emission path, sensing a beacon perimeter emitted by the beacon, and moving along the beam perimeter through the gateway and into the second area.


In another aspect, a method of navigating an autonomous coverage robot between bounded areas includes positioning a navigation beacon in a gateway between adjoining first and second bounded areas. The beacon is configured to transmit a directed emission into the first bounded area and a gateway emission in a cross-gateway direction. The method also includes placing the coverage robot within the first bounded area. The robot autonomously traverses the first bounded area in a cleaning mode and upon encountering the gateway emission, the robot remains in the first bounded area, thereby avoiding the robot migration into the second area. Upon termination of the cleaning mode in the first area, the robot autonomously initiates a migration mode to move through the gateway into the second bounded area by detecting the directed emission and, in response to the directed emission, moves past the beacon through the gateway and into the second area.


The navigation beacon includes a base positionable in the gateway, the base defining a cross-gateway direction. A gateway beam emitter is housed in the base and arranged to emit a beam in the cross-gateway direction. First and second directional beam emitters are housed in the base and arranged to emit respective focused beams into each of the adjoining areas when the base is positioned in the gateway. An omni-directional beam emitter is disposed on the base and configured to project a proximity beam laterally around the beacon. The first and second directional focused beams form angles adjacent the cross-gateway direction of between about 45 and 90 degrees.


Another example of the navigation beacon includes a beam emitter emitting a beam along the line of a “virtual wall” and a circular proximity beam (and/or RF zone) to prevent the robot collisions when detected by the robot. A first beam emitter may be positioned in the gateway to emit a beam across the gateway. A second beam emitter may be positioned in the gateway to emit a beam angled into a first room and modulated to signify the first room. A third beam emitter may be positioned in the gateway to emit a beam angled into a second adjoining room and modulated to signify the second room.


The navigation beacon can also act as a virtual temporary confinement wall or gateway, where the robot stays within the virtual gateway(s) for, e.g., a predetermined number of virtual gateway interactions or time span, then crosses the virtual gateway and resets its time or incident counter to dwell in the next “room” which could be partially or wholly defined by virtual gateways.


Multiple beacons may be employed to define the boundaries of each bounded area. The directed vectoring and gateway emissions may be infrared light. Each transmitted emission is differentiated by time division multiplexing with a fixed period random offset. Each beacon is encoded differently (e.g. left and right beams) and can be set by a DIP switch or other such device located on the beacon. The beacon encoding can be set or changed by the robot (e.g. via RF commands issued by the robot) as the robot encounters successive beacons or two beacons with the same encoding. The beacon may be set to be virtual walls or threshold markers (or both). As the robot encounters each beacon, the robot counts the beacons and identifies them by their modulation. The first one encountered is number one, and is believed to be nearest a base. The second one is number two, farther yet. The third one is number three, farther yet. The left and right side beams or directed beams of the beacons are encoded differently. As a result, the robot can record which side is thought to be nearer to the base. When the robot is ready to return to the base (e.g. timed out, low battery remaining, bin full, schedule finished, maintenance needed, general event triggered, or under any other circumstances), the robot looks for the lowest numbered beacon with an away-side beam and heads in that direction. The directionality of the beam also allows the robot to maneuver past the beacon appropriately. For example, on the way out from the base station, the robot passes directed beam A and directed beam B, sequentially, of a beacon, either one or both on the left side of the robot. To return to the base station the robot should wait/hunt/clean until it detects directed beam B of the beacon. The robot then approaches the beacon by following beam B until it detects a proximity beam of the beacon. The robot turns left and proceeds in a curved path along an edge of the proximity beam while hunting/cleaning until the robot detects beam A. The robot follows beam A out away from the beacon and then hunts/cleans for the base station. This is merely one example of how to navigate using the directionality of the room-identifying left and right beacons.


In some examples, upon termination of the migration mode in the second area, the robot autonomously initiates a docking mode to maneuver towards a base station in the second area and docking with the station. The base station includes a base defining a docking direction from which the robot may properly dock, a robot charger housed in the base, an omni-directional beam emitter mounted on the base and configured to project a proximity beam latently around the base station, and two navigational field emitters housed in the base and arranged to emit respective, laterally bounded and overlapping fields of emissions of signal beams, respectively. One of the emitted fields defines a lateral field edge aligned with the docking direction and overlapped by the other of the fields. The robot maneuvers towards the base station by detecting and advancing along the lateral field edge of the overlapping fields aligned with the docking direction until docked with the station. The two navigational field emitters, which may be infrared emitters, are positioned to form an angle therebetween of between about 45 and 90 degrees. The robot detects the emissions of the base station with an omni-directional receiver on the robot and maneuvers to detect an outer lateral field edge of at least one field emission. The robot advances along the outer lateral field edge of the one field emission to the aligned lateral field edge of the overlapping fields. Upon detecting the aligned lateral field edge, the robot advances along the aligned lateral field edge until docked with the station.





DESCRIPTION OF DRAWINGS


FIG. 1A is a perspective view showing an example of an autonomous coverage robot.



FIG. 1B is an exploded views of an autonomous coverage robot. FIG. 1C is an exploded views of a bumper of an autonomous coverage robot.



FIG. 2 illustrates the positions of omni-directional receiver and directional receiver on the bumper of an autonomous coverage robot.



FIG. 3 is a perspective view of a directional receiver.



FIG. 4 is a front view of a directional receiver.



FIG. 5 is an exploded view of the directional receiver. Like reference symbols in the various drawings indicate like elements.



FIGS. 6-8 show perspective, side, and cut-away views of the omni-directional receiver.



FIG. 9 shows a below-perspective view autonomous coverage robot.



FIG. 10 shows an exploded view of an example navigation beacon.



FIGS. 11 and 12 show perspective and front views of an example beacon emitter assembly.



FIG. 13 shows an example of a simplified navigation beacon that may be used with an autonomous mobile robot navigation system.



FIGS. 14-17 show various views of an example base station that may be included in the autonomous mobile robot system.



FIGS. 18-21 show perspective, top, rear, and side views of an example emitter assembly used in base station.



FIG. 22 shows a block diagram of representative electronic structure of an autonomous coverage robot.



FIG. 23A shows a first example of applying of a method of navigating an autonomous coverage robot between bounded areas using navigation beacons with the addition of a base station; and FIG. 23B shows a second example in which the robot encounters more than one beam at the same time.



FIGS. 24A and 24B are block diagrams illustrating a software architecture for navigating an autonomous coverage robot.



FIGS. 25A-C show schematic views of a robot moving from one area to another using a navigation beacon.



FIG. 26 shows a schematic view of a beam align behavior of a robot.



FIG. 27 shows a schematic view of a directed beam homing behavior of a robot.



FIG. 28 shows a schematic view of a field align behavior of a robot.



FIG. 29 shows a schematic view of a field following behavior of a robot.



FIG. 30 shows a schematic view of a leaving beacon behavior of a robot.



FIGS. 31A-H show schematic views of a robot moving from one area to another using a navigation beacon.



FIGS. 32A-E show schematic views of a robot docking with a base station.





Like reference symbols in the various drawings indicate like elements.


DETAILED DESCRIPTION


FIGS. 1A and 1B show above-perspective and exploded views of an autonomous coverage robot 100. The robot 100 has a chassis 102, a controller 103, an omni-directional receiver 104, and a directional receiver 106. Chassis 102 has a forward drive direction and carries controller 103 and the receivers 104 and 106 on a bumper 107. Receivers 104 and 106 provide navigation information to controller 103. Using input from receivers 104 and 106, controller 103 generates commands to be carried out by the robot 100. As a result, the robot 100 is capable of cleaning floors in an autonomous fashion.



FIG. 2 illustrates the positions of omni-directional receiver 104 and directional receiver 106 on the bumper 107 of the robot 100. Bumper 107 may also have other sensors, as shown in FIG. 1C, to aid the robot 100 as it navigates about its surroundings. Proximity sensors 1072 may be used to determine when an obstacle is close to the robot 100. Contact sensors 1074 may be used to determine when the robot 100 has physically encountered an object. Cliff sensors 1076 may be used to sense when the robot 100 has encountered the edge of the floor, such as when it encounters a set of stairs.


Omni directional sensors 104 may be used to sense when the robot 100 is in close proximity to a navigation beacon. For example, the omni-directional sensor 104 may relay a signal to a control system that indicates the strength of an emission, where a stronger signal indicates closer proximity to a navigation beacon.


Proximity sensors 1072 may be used to detect when an obstacle is nearby. The proximity sensors 1072 may, for example, be infrared or ultrasonic sensors that provide a signal when an object is within a given range of the robot 100.


Contact or bump sensors 1074 may be used to detect if the robot 100 physically encounters an obstacle. Such sensors 1074 may use a physical property such as capacitance or physical displacement within the robot 100 to determine when it has encountered an obstacle.


Cliff sensors 1076 may be used to sense when the robot 100 has encountered the edge of the floor, such as when it encounters a set of stairs. The robot 100 may have behaviors that cause it to take an action, such as changing its direction of travel, when an edge is detected.


In some implementations, cliff sensors 1074 may be installed within a mounting apparatus that stabilizes and protects the sensor and which positions the sensor to point towards the window installed onto the bottom of the mounting apparatus. Together the sensor, the mounting apparatus and the window comprise a cliff sensor unit. There may, for instance, be four cliff sensor units installed within the bumper.


Reliability of a cliff sensor may be increased by reducing dust buildup.


In some implementations, a window may be installed on the bottom of the mounting apparatus which includes a shield mounted within a slanted molding composed of a material which prevents dust build up, such as an antistatic material. The shield component and the molding may be welded together. To further facilitate the reduction in dust and dirt buildup, the shield may be mounted on a slant to allow dirt to more easily slide off.


In some implementations, a secondary cliff sensor may be present behind existing cliff sensors to detect floor edges in the event that a wheel drop sensor on the caster wheel fail.


Controller 103 may be configured to propel the robot 100 according to a heading setting and a speed setting. Signals received from proximity and contact sensors may be used by the control system to issue commands that deal with obstacles. Signals from the proximity and contact sensors may cause the control system to change the commanded speed or heading of the robot 100. For instance, a signal from a proximity sensor due to a nearby wall may result in the control system issuing a command to slow down. In another instance, a collision signal from the contact sensors due to an encounter with a chair may cause the control system to issue a command to change heading. In other instances, the speed setting of the mobile the robot may not be reduced in response to the contact sensor; and/or the heading setting of the mobile the robot may not be altered in response to the proximity sensor, but instead the two sensors may operate independently.


Controller 103 may include a first independent behavioral routine configured to adjust the speed setting of the mobile the robot; and a second independent behavioral routine configured to alter the heading setting of the mobile the robot, in which the first and second independent behavioral routines are configured to execute concurrently and mutually independently. The first independent behavioral routine may be configured to poll the proximity sensor, and the second independent behavioral routine may be configured to poll the kinetic bump sensor.



FIG. 3-5 show perspective, front, and exploded views of the directional receiver 104. The directional receiver 104 is positioned on the front of the chassis 102. Emissions may be received along the drive direction by the directional receiver 104 which generates and sends corresponding signals to controller 103. The directional receiver 104 includes a mounting receptacle 108 with a first aperture 110 and a second aperture 112 that are aligned with the drive direction of the chassis 102. Associated with the apertures 110, 112 are a first component receiver 114 and a second component receiver 116.


The component receivers 114, 116 are positioned relative to the apertures 110, 112 such that emissions which originate along the driving direction and fall on the directional receiver 104 may pass through the apertures I 10, 112 and onto their respective component receivers 114, 116. Emissions falling on the component receivers 114, 116 result in a corresponding signals being generated which may be used by controller 103 to determine the direction of the emission relative to the robot's drive direction. A cover 117 attached to the top of the mounting receptacle 108 to prevent emissions that do not originate along the driving direction from falling on the component receivers 114, 116. In some implementations, locking tabs 119 or some other fastening method, may be used to secure the cover 117 to the mounting receptacle 108.


In some implementations, the component receivers 114, 116 of the directional receiver 104 can detect a homing beam with an overlap of substantially between 10% and 30%. In one example, the first and second component receivers 114, 116 may be positioned such that emissions falling squarely on each would form an angle in the range of 1 to 15 degrees. In another example, the first and second component receivers 114, 116 may be aligned so emissions aligned with them cross at a point 3 to 5 meters in front of the robot. In one example, each component receiver 114, 116 receives emissions within a sweep angle of between about 28 to 33 degrees with a center of the sweep at about 14 degrees from a direction normal to the directional receiver 104 and a sweep overlap of about 10 degrees with the other component receiver 114, 116.


In some cases, the controller 103 may maneuver the robot 100 to detect an emission with the omni-directional receiver 104. The direction of the emission may be determined using the component receivers 114, 116. In response to the detection, the controller 103 may orient the chassis 102 to align the robot drive direction with the determined emission direction.


In other cases, controller 103 may orient the chassis 102 to align the robot drive direction with the determined emission direction based on the signals generated by the component receivers 114, 116 in response to emissions detected with the directional receiver 106.



FIGS. 6-8 show perspective, side, and cut-away views of the omni-directional receiver 104. The omni-directional receiver 104 may include a housing 118, a conical reflector 120 and an emission receiver 122. The housing 118 has an upper portion 124 and an inner cavity 126. The upper portion 124 may allow a transmission of an emission into inner cavity 126. Conical reflector 120 is located on an upper surface of the cavity 126 to reflect emissions falling on the upper portion 124 of the housing 118 into the inner cavity 126. Emission receiver 122 is located in inner cavity 126 below conical reflector 120.


In some implementations, the receivers 114, 116, and 122 may be configured to receive transmissions of infrared light (IR). In such cases, a guide 128 (e.g. a light pipe) may guide emissions reflected off the conical reflector 120 and channels them to emission receiver 122.



FIG. 9 shows a below-perspective view of the autonomous coverage robot 100. A drive system 130 includes a first drive wheel 132 and a second drive wheel 134 which support chassis 102. A castor 136 may provide additional support to chassis 102. Motors may be mechanically coupled to the drive wheels to propel the robot 100, providing forward, reverse, and turning capabilities.


The robot 100 may have a floor cleaning system that removes dirt and debris from the floor as it maneuvers about. The floor cleaning system may include a floor cleaning assembly 140, a cleaning bin 142 and a bin sensor (not shown). The floor cleaning assemblies 140 and 146 and cleaning bin 142 may be carried by the chassis 102. Cleaning bin 142 may be arranged to collect debris removed from a floor being cleaned by cleaning assembly 140. The bin sensor may be configured to generate a signal that indicates whether the cleaning bin has been filled to a threshold. When the threshold has been reached, controller 103 may initiate a docking sequence with a base station (described below). Cleaning bin 142 may be accessed to empty its contents either manually or, in some implementations, the robot 100 may automatically empty cleaning, bin 142 when docked.


In addition to providing a framework for the other elements that make up the robot 100, the chassis 102 may have a ridge on its front which is higher than all other points on the chassis 102. Such a ridge may be stop the robot 100 if it encounters an overhead obstacle, such as a couch, and prevent it from becoming wedged underneath. To enable the robot 100 to more easily free itself in the event it becomes wedged, controller 103 may normally command the drive system 130 to operate at less than maximum torque, for instance at 50% of capacity. If the robot 100 becomes wedged, for instance, sensed by increased current flowing to the drive motors, controller 103 may command increased torque to free the robot 100.


Another example of an anti-wedging system includes a spring loaded wheel system having a potentiometer to measure how much the robot 100 is pushed down by a potential wedging obstacle. Another example includes an infrared sensor disposed on a lower surface of the robot 100, where the infrared sensor is used to measure a distance that the robot 100 is pushed down.



FIG. 10 shows an exploded view of an example navigation beacon 150. Navigation beacon 150 may be used in conjunction with existing walls or other obstacles to create a bounded area. Bounding an area may be done, for example, to restrict a robot from entering or leaving an area. The navigation beacon 150, along with the robot 100 described above, form an example of an autonomous mobile robot system.


Navigation beacon 150 has a beacon emitter assembly 152, which includes a gateway beacon emitter 154 and an omni-directional emitter 160. Gateway beacon emitter 154 may be arranged to transmit a gateway marking emission. For instance, beacon 150 may be located within a gateway that separates a first and second adjacent areas and emit the gateway marking emission to form a boundary. Omni-directional receiver 104 and directional receiver 106 on the robot 100 may detect the gateway marking emissions and thereby function as beacon emission sensors. For example, beacon 150 and the robot 100 may use infrared light (IR) emitters and sensors to create and detect the gateway marking emissions.


In one example, the robot 100 controls the state of the navigation beams through commands transmitted over a packet radio network. The address that the beacons respond to on this network is a combination of a robot address and a node address. After installing batteries in the beacon 150, the beacon 150 periodically tries to contact any robot to see if it should wake up and operate its emitters. A robot 100 may respond by transmitting a radio packet containing an invitation to join its network and a temporary address to use. While operating with a temporary address, the beacon 150 transmits an infrared code in a fence beam from emitter 154 and force field beam from omni-directional emitter 160 indicating that it is not bound, i.e. its radio address is temporary. When a robot 100 sees a beam with the bound code, it iteratively transmits a radio packet to every recently assigned temporary address to send a new code in its beam called a wink. If the robot 100 sees a wink code, it transmits a radio packet containing a new node address to be used from now on as well as a time indicating the number of hours the address is valid for use. Having bound to a robot 100 successfully, the beacon 150 will henceforth only wake up in the presence of that robot 100. The technique of confirming radio communications using infrared light signals is designed to prevent a robot 100 that is not on the same floor as the beacon 150 from controlling it permanently.


Drive system 130 may be configured to maneuver the robot 100 about the first area while in a cleaning mode. In the cleaning mode, the robot 100 may be redirected in response to detecting the gateway marking emission. In addition, the drive system 130 may also be configured to maneuver the robot 100 through the gateway into the second bounded area while in a migration mode.


While in the cleaning mode, the robot 100 may remain in the first bounded area by changing its drive direction when it detects the gateway marking emission. The gateway marking emission thus acts as a virtual barrier which may prevent the robot 100 from leaving the first bounded area.


In some instances, the drive system 130 may be configured to maneuver the robot 100 about the first bounded area in the cleaning mode for a preset time interval. When the preset time interval elapses, the drive system 130 may be automatically put in migration mode. While in migration mode, drive system 130 may be configured to maneuver the robot 100 across the gateway emission. In other instances, the drive system may configured to maneuver the robot about the first bounded area in the cleaning mode until it encounters of the gateway marking emission a preset number of times. Once the gateway marking emission has been encountered the preset number of times, the migration mode may be automatically initiated.


The robot 100 may include a transmitter for communicating with beacon 150. The transmitter may be used to signals beacon 150 to halt or pause transmission of the gateway emission in the migration mode. By signaling beacon 150 to turn on its various emitters only when needed, the system may implement a power-saving function. Such a function may serve to extend battery life in beacon 150.



FIGS. 11 and 12 show perspective and front views of beacon emitter assembly 152. Beacon emitter assembly 152 includes a first directed vectoring beacon emitter 156 and a second directed vectoring beacon emitter 158. Directed vectoring beam emitters may be used to create an emission field with a characteristic edge and spread pattern that may be used to define a navigation route. Navigation beacon 150 may be located within a gateway between two bounded areas with vectoring beacon emitter 156 arranged to transmit a directed vectoring emission into the first bounded area. The angle between the directed vectoring emission and the gateway may be, for example, in the range of 45-90 degrees. In some instances, the directed vectoring emission may consist of infrared light.


While in migration mode, drive system 130 may be configured to direct 30 the robot 100 toward beacon 150 when it encounters the directed vectoring emission emanating from vectoring beacon emitter 156. The robot 100 may then determine the position of the beacon 150 based on the detected direction of the directed vectoring emission relative to directional receiver 106. Once position is determined, the robot 100 may align itself in a drive direction relative to the directed vectoring emission. For example, the robot 100 may advance forward along the path of the directed vectoring emission to reach and traverse the gateway in which beacon 150 is located.


The robot 100 may be able to remotely activate and deactivate the beacon emissions. For example, the robot 100 may use wireless communication, such as radiofrequency (RF) communication, to pass activation and deactivation signals. The robot 100 may remotely activate the directed vectoring emission of the beacon 150 and deactivate the gateway emission upon initiating the migration mode. The robot 100 may remotely deactivate the directed vectoring emission of beacon 150 and activate the gateway emission upon terminating the migration mode.


In some instances, the beacon 150 may be configured to communicate schedule information with the robot 100. For example, beacon 150 may be able to transmit times for which the robot 100 should enter cleaning mode, when it should enter migration mode, etc. Schedule information may include details such as starting or ending times and dates.


Navigation beacon 150 may also transmit a proximity emission laterally about itself. The robot 100 may avoid cleaning and migration within the proximity emission by executing a pre-configured behavior such as changing its course when the proximity emission is detected. The proximity emission may be thought of as a “force field” through which the robot 100 is not permitted to pass.


Navigation beacon 150 may have a switch to select between a confinement mode, a navigation mode, and an off mode. Navigation beacon 150 may have a range select switch like a virtual wall. There may be a light to indicate when the navigation beacon is operating. The navigation beacon 150 may turn on an operational indicator light it is enabled or only when commanded to do so by the robot 100. There may also be a low battery warning, and there may be a separate low battery light.


In order to reduce power consumption and thus increase battery life, different operational modes may exist for the navigation beacon. When the robot 100 is not operating, navigation beacon 150 may be in a low power mode in which emitters are off and the navigation beacon periodically monitors the communication link to determine if a wakeup is necessary. When the robot 100 is operating, it may send a signal to a single or multiple navigation beacons to turn on their respective emitters. There may be different commands to turn on the various emitters. In addition, the navigation beacon may return to low power mode after an elapsed time.


In the event multiple navigation beacon are used with the robot 100, each navigation beacon may have a unique identifier, such as a 16-bit (or larger) identification number contained in memory. This identifier may be generated in the factory or by the navigation beacon itself. If the identifier is generated in the factory it may be stored in non-volatile memory in the beacon. The robot 100 may use the unique identifier to generate an internal map which may be used to navigate from one bounded region to another by navigating from one navigation beacon to another. In some implementations, the identifier for navigation beacon 150 may be used to generate signals encoded in its emissions that may be detected by the robot 100. In the event that navigation beacon 150 is unable to transmit the entire identification number, a unique code maybe generated as a derivation of the ID or by negotiation between the navigation beacons that are present and/or the robot 100.


The robot and navigation beacons may communicate via point to point or via broadcast transmissions. In the point to point scheme, the robot would have to learn the identifiers of all of the navigation beacons. Then it could send transmissions with a data field indicating the destination navigation beacon. That navigation beacon and only that navigation beacon would respond accordingly. This has the advantage that only specific navigation beacons would respond, keeping other navigation beacons off and thus increasing their battery life. User interaction may be used to train the robot to know the navigation beacons.


An alternate concept is for the robot to broadcast to all navigation beacons a command to enable them. This scheme works without user interaction. In order to mitigate this effect of increased battery consumption and reduced battery life on these navigation beacons, an alleviating technique is to reduce the power of the robot transmission.



FIG. 13 shows an example of a simplified navigation beacon 152 that may be used with an autonomous mobile robot navigation system. In this example, the beacon has a beacon emitter assembly 152 including a gateway emitter 154 and an omni-directional emitter 160, but does not have vectoring emitters. Such a beacon may be used to define boundaries of a region but not necessarily support migration functionality.



FIGS. 14-17 show various views of an example base station 180 that may be included in the autonomous mobile robot system. Base station 180 may be located in a bounded area served by the system. Base station 180 has a base 182 with a defined docking direction and a robot charger 184. The robot 100 may approach base 182 from the docking direction to dock with the base 182. The robot charger 184 is housed in the base 182 and may be used to charge the robot 100 when it is properly docked.


The robot charger 184 may begins operating when triggered by a detection circuit after the robot 100 is docked to base 182. Battery conditions may be used to govern whether a deeply discharge mode, a pre-charge trickle mode, or a post charge trickle mode is used to charge the battery.



FIGS. 18-21 show perspective, top, rear, and side views of an example emitter assembly 186 used in base station 180. Emitter assembly 186 has a lower housing 188 which holds an omni-directional beam emitter 190 and two navigational field emitters 192 and 194.


The omni-directional emitter 190 may be mounted on housing 188 and be configured to project a proximity beam laterally around the docking station through the use of an upper housing 196. Upper housing 196 may be configured such that emissions from omni-directional emitter 190 form a desired pattern outside base 182, such as a circle.


Navigational field emitters 192 and 194 may be attached to lower housing 188 and arranged to emit respective, laterally bounded and overlapping fields of emissions of signal beams. One of the emitted fields may define a lateral field edge aligned with the docking direction and be overlapped by the other emitted fields. The two navigational field emitters of the base station may be positioned such that their angle 30 of overlap may be about 6 degrees and each emitter's aperture angle may be 20-30 degrees. A projection 198 on lower housing 188 may be used to shape the emissions from emitters 192, 194 to achieve the above pattern. In one example, the navigational field emitters 192 projects a beam at about 12 degrees from a direction normal to the emitter assembly 186, where the beam sweeps about 40 degrees starting from about −5 degrees to about 35. The other navigational field emitters 194, with the L-shaped baffle or mask 198 terminating near the middle of the LED, projects a beam at about 12 degrees from a direction normal to the emitter assembly 186, where the beam sweeps from straight out to about −35 degrees. In some implementations, the emission fields may consist of infrared light (IR). In such cases, field emitters 190, 192, and 194 may consist of infrared light emitting diodes (LEDs).



FIG. 22 shows a schematic view of electronics 101 of the robot 100. The electronics 101 include the controller 103 which communicates with a bumper micro-controller 107A that controls the omni-directional receiver 104, directional receiver 106, wall proximity sensors 1072, and bumper switch 1074. The controller 103 monitors all other sensor inputs, including the cliff sensors 1076.



FIGS. 23A-B show an example of applying of a method of navigating an autonomous coverage robot between bounded areas using navigation beacons. The method may be applied to a system such as the example the robot navigation system described above. While the following description relates a particular software architecture and set of algorithms to the navigation beacons disclosed as a physical infrastructure, the very same embodiments of a robot, control and software architecture, and/or many particular algorithms can be used with a different physical infrastructure.


Accordingly, because the manner and methods by which the robot handles sequences of goals and navigation are relatively independent of the actual physical form of the infrastructure, for discussions involving such manner and methods, “navigation beacon” includes active and passive landmarks of other physical kinds (e.g., bar codes, retro-reflective tags, illuminated or IR spots projected on the ceiling or walls, directional RF or visible light, and other features identifiable in the environment), and “beam” includes directed vectors of other physical kinds (including inferred directional vectors calculated or kept in memory that may be followed, as well as line-of-sight directions to distant observed objects). In the example shown, walls 201, a first navigation beacon 202 and a second navigation beacon 203 (also discussed herein as “lighthouses”) define a first bounded area 204, a second bounded area 206, and a third bounded area 207 (which are typically rooms of a home). Navigation beacon 202 may be configured to transmit a gateway marking emission 208 (also noted herein as a “virtual gateway” or “gateway beam”) across a gateway 210 (typically a doorway, but openings of a few feet to 10 or 12 feet are possible, as well as arbitrary segmenting of a room without any particular openings). The gateway marking emission 208 may be an IR signal, for example, which can be sensed by a coverage robot 212.


The robot 212 may be placed within a bounded area 204. The robot 212 may autonomously traverse the bounded area 204 in a cleaning mode. If the robot 212 encounters gateway marking emission 208 in gateway 210, the robot 212 remains in the bounded area 204 without moving into an adjacent bounded area 206. Upon termination of the cleaning or coverage mode in the first area, the robot 212 may autonomously (i.e., in response to an internal trigger condition as discussed herein, e.g., time expired, distance covered, percent covered) enter a migration mode in which the robot continues cleaning but monitors for the gateway 210, locates a gateway 210, moves to and through gateway 210, (past beacon 202,) into the adjacent bounded area 206.


In some implementations, the robot 212 may autonomously enter the cleaning mode after the migration to bounded area 206 has been completed. While in the migration mode, the robot 202 may be non-responsive to gateway emission 208 (although the robot may record having crossed the gateway emission 208). In some cases, the robot 212 may remotely (i.e., by direct or relayed RF communications or line-of-sight or reflected optical signaling) deactivate gateway marking emission 208 upon entering migration mode. The robot 212 may also activate gateway marking emission 208 after exiting the migration mode.


Furthermore, navigation beacon 202 may transmit a proximity emission or field 214 laterally around (surrounding) itself, establishing an exclusion zone or “force field”. The robot 212 may detect this proximity field and use the detection to attempt to avoid maneuvering within the exclusion zone around a beacon 202 while in the cleaning or migrating modes. The primary purpose of the exclusion zone is to prevent the robot 202 from striking the beacon 202 itself and moving it.


Navigation beacon 202 may be configured to transmit a first directed vectoring emission 216 (also discussed herein as a “directed emission,” “directed beam,” “launch beam” or “tractor beam”) into bounded area 204 and a second directed vectoring emission 217 into bounded area 206, with beacon 202 positioned within gateway 210. For example, the robot 212 may drive toward beacon 202 upon detecting directed emission 216 while in the migration mode.


While in the migration mode, the robot 212 may detect the directed emission 216 by sensing the directed emission 216 with a directional receiver 218 (also discussed herein as a “binocular sensor,” e.g., including twin detectors in collimated tubes) located on the front of the robot 212 (i.e. the side of the robot 212 that is aligned with the forward the robot drive direction). Alternatively, the robot 212 may sense the directed emission 216 with an omni-directional receiver 222 prior to detection by directional receiver 218. The term “omni-directional receiver” as discussed herein includes non-directional and multi-directional receivers, e.g., receivers including two or more sensors (at compass points, etc.) directed about the periphery of the robot. The robot 212 may use the signal detected by omni-directional receiver 222 to then maneuver (turn in place or turn going in forward or reverse directions) to sense directed emission 216 with directional receiver 218. The robot 212 may align its direction of travel with a path defined by an edge 220 of directed emission 216. Directional and omni-directional receivers 218 and 222 may have construction and function similar to that described above.


In some cases, the robot 212 may move towards the beacon 202 along the emission path (i.e., servoing using the two beam detectors 114 and 116 of the binocular sensor 106) until it senses a beacon perimeter 224 (i.e. the edge of proximity field emission 214). The robot 212 may move along beam perimeter 224 through the gateway 210 and into bounded area 206. The angle between the gateway emission 208 and directed emission 216 may, in some instances, be between about 45 and 90 degrees, and optionally is about 60 degrees. The gateway emission 208 is sized according to the robot diameter, and diverges to be approximately the width of the robot at the proximity field edge or range of detection (e.g., to prevent the robot crossing at that point). This may be a span of 10 degrees or more, but is determined by the robot diameter. The angle between the center of the gateway emission 208 and the center of each directed emission 216 is, in two different examples, about 45 degrees or about 60 degrees, and each directed emission 216 beam is a diverging beam made diffuse by the slot mask near the emitter, and is, in two different examples, about 20-30 degrees (e.g., 25 degrees) or about 30-50 degrees (e.g., 40 degrees). In some instances, the directed emissions 216 and 217 may consist of infrared light.


The robot 212 may remotely activate and/or deactivate emissions coming from beacon 202, such as directed emission 216 or gateway emission 208. Various methods may be used to transmit signals between the robot 212 and the navigation beacons 202 and 203.


In particular, it is effective to use the same inexpensive, common IR-LED emitters for each of the emitters of each of the beacons 202, 203, and in addition for a recharging or other dock 240 for the robot 212. The same sensors on the robot 212 can detect all of the emitters; and different sensors on the robot 212 (e.g., the multi-directional as well as the collimated directional sensors) can detect the same emitters for different purposes (as discussed herein, e.g., following, homing, stopping). In order to differentiate between the different emitters, each emitter may be encoded (e.g., modulated with a different serial code). This also helps avoid confusion with household and other remote controls as well as IR components of sunlight and other ambient sources. However, as shown in FIG. 23, the robot 212 can encounter situations in which it is traveling through the emission paths of several different emitters.


In this situation, especially when using optical multi-directional receivers or a common optical omni-directional receiver (but also in other instances, e.g., RF) the sensor or software structure is configured to be capable of discerning one emitter from another when several signals are simultaneously received. If different frequencies are available, then these may be employed. One strategy is for the emitters to take turns, being synchronized by their communications with the robot or each other. However, this is complex and introduces a failure mode when synchronization breaks down. It is advantageous for each beacon 202, 203 and dock 240 to self-control, and stagger their own signals in time, with intervals between each emission. For instance, each transmitted emission may be differentiated by time division multiplexing (i.e., multiplexing as the beam origins are taken together as a system, including within a beacon having more than one beam or among several beacons) with a fixed period random offset (e.g., which may be different for each beam source). The interval or random offset may be changed from time to time (e.g., at random intervals) or by communication with the robot (e.g., by RF communication when the robot detects an interference condition). At manufacture, or upon changing offset, e.g., the offset may be selected from a set of offsets unlikely to harmonically interfere, or that do not share common factors. In this manner, the robot may encounter several emission paths at once yet discern the identity of each. In other instances, each transmitted emission may be differentiated by different wavelengths of light, infrared modulations, and wavelength filters/windows on the emitters and receivers.


The robot 212 may use RF communication to signal to beacon 202 to remotely activate directed emission 216 and deactivate gateway emission 208 upon initiating the migration mode. In another example, the robot 212 may remotely deactivate the directed vectoring emission 216 and activate gateway emission 208 upon exiting the migration mode.


In some instances, the robot 212 may activate a gateway emission 226 separating bounded areas 206 and 207 and initiate the cleaning mode. Similarly to that described above, the robot 212 may be prevented from leaving bounded area 206 when it encounters gateway emissions 208 and 226. When finished cleaning in bounded area 206, the robot 212 may initiate the migration mode and navigate through gateway 228 by following directed emission 230 to a perimeter 234 of a proximity emission 232 and enter bounded area 207. Once in bounded area 207, the robot 212 may re-enter the cleaning or working mode. After, for instance, a set time period or preset number of encounters with gateway emission 208, the robot 212 may migrate from bounded area 204 to bounded area 206.


Alternatively, a virtual wall emitter (gateway) may be set to independently deactivate itself on a schedule to permit the robot to cross the gateway, e.g., acting as a virtual wall for a first interval (to confine in a first or current room), then deactivating temporarily or intermittently, then acting as a virtual wall for a second interval (to confine in a second or successive room). This technique may also be interactive, e.g., the robot may use RF communication with one or more beacons to deactivate a virtual wall or bypass/cross a virtual wall.


It should be noted that in all instances herein, the robot's behavior system is arranged such that obstacle events, including cliff detection events, have a higher priority than any room-to-room navigation or cleaning. Accordingly, for example, when the robot encounters a cliff detection in the middle of a navigation or other set of sequenced behaviors, the robot may nonetheless avoid the cliff (aborting the current part of the sequence and resetting the state of the sequence).



FIG. 23 shows a set of rooms similar to FIG. 22, but arranged (with the addition of a base station) to illustrate the situation in which the robot encounters multiple beams from two beacon emitters and also from a dock 240. Using the strategies avoiding beam confusion noted above, the robot 212 may nonetheless navigate from room to room or the robot 212 may autonomously initiate a docking mode to maneuver towards the base station 240 in bounded area 206 and dock with the station upon termination of the migration mode in the second area.


The base station 240 may include a base, a robot charger, an omni-directional beam emitter and two navigational field emitters and be similar to the base station 180 described above. The robot 212 may maneuver towards base station 240 by detecting and advancing along one of the lateral field edges 242 or 244 of the overlapping fields 246, 248 aligned with the docking direction until docked with the station 240.


The robot 212 may detect the emissions of base station 240 with omni-directional receiver 222 on the robot and maneuver to detect an outer lateral field edge (e.g. 247) of at least one field emission 246 or 248. The robot 212 may then advance along outer lateral field edge 247 or 249 to the aligned lateral field edge 242 or 244 of the overlapping fields. Upon detecting the aligned lateral field edge 242 or 244, the robot 212 advances along the aligned lateral field edge 242 or 244 until docked with base station 240.



FIG. 24A is a block diagram showing a behavioral software architecture within a controller 250, in which goal-oriented or sequenced behaviors are depicted in simplified form. FIG. 24B shows the same software architecture, but provides more detail on the goal-oriented or sequenced behaviors while simplifying some of the elements of FIG. 24A.


While embodiments of the robot discussed herein may use behavioral based control only in part or not at all, behavior based control is effective at controlling 5 the robot to be robust (i.e. not getting stuck or failing) as well as safe. Accordingly, FIGS. 24A and 24B depict a software architecture that permits improved control of a mobile robot. Rather than being a purely reactive architecture, FIGS. 24A and 24B introduce an element of planning that is partially reactive (and therefore more robust)—goal oriented, sequenced reactive behaviors.


In the reactive yet sequenced behaviors, the implied success of a related behavior in a sequence is an enabling condition (among others) for each behavior beyond the initiating behavior in the sequence (some sequences may have more than one initiating behavior or branches). The implied success is kept as a state signifying an enabling condition of a successive behavior, in the form of a flag, state machine, or the like. The state may be set or changed by a behavior itself upon completion, by an arbiter process, or by a planning process.


Referring to FIG. 24A, the robot 302 employs a control and software architecture that has a number of behaviors that are executed by an arbiter 252 in controller 250. A behavior is entered into the arbiter 252 in response to a sensor event, or a sequence state event from the planner 253. In one embodiment, all behaviors have a fixed relative priority with respect to one another. The arbiter 252 (in this case) recognizes enabling conditions, which behaviors have a full set of enabling conditions, and selects the behavior having the highest priority among those that have fulfilled enabling conditions. The diagram shown in FIGS. 24A and 24B does not necessarily reflect the (fixed) priority hierarchy of the robot 302. In order of decreasing priority, the behaviors are generally categorized as escape and/or avoidance behaviors (such as avoiding a cliff or escaping a corner), working behaviors (e.g., wall following, bouncing, or driving in a straight line), and for the purposes of this application, sequenced reactive behaviors. The behaviors may include: different escape behaviors 255, 255a, 255n (including escaping corners, anti-canyoning, stuck situations, “ballistic” temporary fire-and-forget movement that suppress some avoid behaviors, e.g., as disclosed in U.S. Pat. No. 6,809,490, the entire contents of the which are hereby incorporated by reference) cliff avoiding 256, virtual wall avoiding 258 (a virtual wall may be a beacon with a gateway beam), other avoidance 258n as a class (e.g., avoiding people via a pyrometer, in general the class of detecting and moving away from something that is a hazard to the robot 302 or to which the robot 302 may be a hazard), spot coverage 264a (covering in a confined pattern such as a spiral or boustrophedon patch), align 264b (turning in place, using side proximity sensors to align with a forward obstacle encountered while obstacle following, e.g., an inside corner), following 264c (representing either or both of substantially parallel or bump following along an obstacle using a side proximity sensor or bumper that extends to the side of the robot), responding to a bump in order to “bounce” 264d (a behavior that occurs after the robot bumps an object), and drive 266. Movement of the robot 302, if any, occurs while a behavior is arbitrated. If more than one behavior is in the arbiter 252, the behavior with a higher priority is executed, as long as any corresponding required conditions are met. For example, the cliff avoiding behavior 256 will not be executed unless a cliff has been detected by a cliff detection sensor, but execution of the cliff avoiding behavior 256 always takes precedence over the execution of other behaviors that also have satisfied enabling conditions.


The reactive behaviors have, as their enabling conditions or triggers, various sensors and detections of phenomena, but, in general, not (arbitrary) states of a sequence. As shown in FIG. 24A, these include sensors for obstacle avoidance and detection, such as forward proximity detection (multiple), forward bump detection (multiple), cliff sensors (multiple), detection of a virtual wall signal (which may instead be considered a coverage trigger). Sensors of these types are be monitored and conditioned by filters, conditioning 292c, and their drivers 292b, which can generate the enabling conditions as well as record data that helps the behavior act predictably and on all available information (e.g., conversion to one-bit “true/false” signals, recording of likely angle of impact or incidence based on strength or time differences from a group of sensors, or historical, averaging, frequency, or variance information).


Actual physical sensors may be represented in the architecture by “virtual” sensors synthesized from the conditioning and drivers. Additional “virtual” sensors that are synthesized from detectable or interpreted physical properties, proprioceptive or interpreted upon the robot 302, such as over-current of a motor, stasis or stuck condition of the robot 302 (by monitoring a lack of odometry reading from a wheel encoder or counter), battery charge state via coulometry, and other virtual sensors “virtual N.”


In addition, reactive behaviors can act according to enabling conditions that represent detected phenomena to be sought or followed. A beam or wireless (RF, acoustic) signal can be detected without direction; or in some cases with direction. A remote beam or marker (bar code, retro-reflective, distinctive, fiducial, or natural recognized by vision landmark) giving a direction can permit homing or relative movement; without direction the robot 302 can nonetheless move to servo on the presence, absence, and/or relative strength of a detected signal. The reflection of a beam from the robot 302, edge, or line can be similarly detected, and following behaviors (such as obstacle following by the robot 302) conducted by servoing on such signal. A debris or artifact signal can be collected by monitoring debris or objects collected by or traversed by the robot, and that signal can be an enabling condition for a reactive behavior controlling a spot coverage pattern. A general class of “seek/servo N” signals are thus also reactive behavior enabling conditions or triggers.


The robot 302 maintains concurrent processes, “parallel” processes that are not generally considered reactive behaviors. A scheduler 292d may be necessary to allocate processor time to most other processes, e.g., including the arbiter and behaviors, in a co-operative or other multitasking manner. If more threading is available, less processes may be managed by the scheduler 292d. As noted, filters and conditioning 292c and drivers 292b, can interpret and translate raw signals. These processes are not considered reactive behaviors, and exercise no direct control over the motor drives or other actuators. In addition, in the present embodiment, brush motor controller(s) 292a control the main and side brushes, although these may alternatively be controlled by dedicated brush behaviors and a brush control arbiter.


With reference to FIG. 24B (the same representation as FIG. 24A, with other areas of the drawing emphasized/deemphasized for clarity), a special concurrent process is a sequence planner 253. The sequence planner 253 is capable of monitoring and/or maintaining state in order to control sequenced reactive behaviors, e.g., setting enabling conditions (that will be caught by the arbiter 252), and/or resetting states in the case of abort conditions (e.g., time-outs, escape or avoidance behavior execution), and/or monitoring or supervising finite state machines that themselves transition between states. It should be noted that enabling condition states or flags or supervised state machines that request a transition between sequential reactive behaviors can be set by a behavior itself, by the arbiter, or by the sequence planner 253, depending on efficiencies and circumstances. It should also be noted that the sequence planner 253 may optionally be capable of directly controlling the arbiter 252 (or second arbiter 252A) to prefer (e.g., re-prioritize) a behavior or directly control a sequenced reactive behavior without setting an enabling condition, state variable or state machine.


While maintaining state is recognized to potentially reduce the robustness of competent reactive control architecture, sequences of intended actions may require the maintenance of some state. As discussed herein, if sequenced reactive behaviors are controlled to execute using request states which may be robustly reset or erased without primary function failure, the reduction in robustness can be reduced.


One exemplary sequenced reactive behavior set is employed by the robot 302 to transition between rooms or navigate to a distal room. As depicted in FIG. 24B, a homing beam align behavior 294a includes an enabling condition that an align start (e.g., a homing request) enabling condition be activated, and that the beam to be homed upon is detected by a multi-directional sensor. As will be discussed below, the homing request enabling condition can be set and maintained in different ways. A homing beam homing behavior 294b is to execute once alignment is successful. However, because alignment may happen serendipitously (the robot may already be aligned when the homing request is activated), the homing beam homing condition also has as an enabling condition the homing request (flag or state), as well as a detection using a directional sensor (binoculars) that substantially necessarily directed at the beam identified as one to be homed upon (or, e.g., filtered to be equivalent to such a detection). Accordingly, two sequenced reactive behaviors are each separate entry points to the sequence, and share a common reactive sequence request enabling condition. When a sensor-based enabling condition fails, these behaviors are aborted (internally or by the planner 253) and the sequence request remains active (or may be reset). In this case, the sequence may be restarted the next time either sensor-based enabling condition occurs. Optionally, a random component (such as a random distance or time return to random coverage, while disabling the sequence request or homing request enabling condition) is permitted to occur following an abort of a sequential reactive behavior in order to decrease the likelihood of systematic failure or looping.


After the homing beam homing behavior 294b has completed, the sequence is in a state where an assumption is made regarding the state that the robot is located appropriately to align with the field to permit begin field following about a beacon to transition to a new chamber. A (new) align start enabling condition (e.g., field align state or request) may be set by the completing sequenced reactive behavior or by the planner, via condition flag or state machine. In order to accommodate the possibility that the field may be the first detectable signal encountered upon initiation of the homing request, the detection of the proximity field by the multi-directional receiver is an equivalent (OR) condition to the field align request, if the homing request is active when the field is encountered, this combination also should initiate field align behavior 294c. If there are no abort conditions, the field align behavior 294c executes. After the field align behavior 294c has completed, the sequence (of sequenced reactive behaviors) is in a state where a next assumption is made regarding the state—that the robot is oriented appropriately to begin field following. A (new) follow start enabling condition (e.g., field follow state or request) may be set by the completing sequenced reactive behavior or by the planner, via condition flag or state machine. Similarly, the launch behavior 294e may have as a primary enabling condition a state-based request rather than a sensor detection. Each of these behaviors 294a, 294b, 294c, 294d, 294e is detailed herein below.


In two of these cases, optionally, a state to enter the entire sequence (homing request) is maintained and an additional state to begin a next part of the sequence (field follow request; launch request) is set and both conditions are necessary to start the next sequenced reactive behavior. This would permit restarting the sequence immediately after an abort (by maintaining the entire sequence request state). Alternatively, the entire sequence entry request is reset and the only the next sequenced reactive behavior entry condition is set (field align request or field follow request). This would cause the entire sequence to be otherwise restarted (e.g., by a new or continued occurrence of conditions that would cause the planner to set the sequence request) after an abort, which could improve general robustness (although not necessarily goal-specific robustness) by erasing essentially all states.


In two cases, no sensor-based condition is an enabling condition for the next sequenced reactive behavior in the sequence, but an enabling condition is a sequence-based or a state-based request—e.g., field follow request, launch request. Accordingly, a sequence may include sequenced reactive behaviors that include only sequence states or progress as enabling conditions for the next sequenced reactive behavior (such as the field follow behavior, or launch behavior), as well as sequenced reactive behaviors that include both sensor-based conditions and sequence states, progress or requests as enabling conditions (such as the beam align or tractor homing behaviors).


In one case, a state to enter the entire sequence (homing request) is maintained and either a sensor-based condition or a sequence-based request is an enabling condition for a next sequenced reactive behavior (field align) in the sequence -e.g., a homing request AND (field align request OR field detection) is interpreted by the arbiter to permit field aligning if no reactive behaviors have higher priority. Accordingly, a sequence may include sequenced reactive behaviors that include either a combination of sequence state/requests, or both sensor-based conditions and sequence request/states, as enabling conditions for a next sequenced reactive behavior.


It should be noted that entry and mid-sequence branching is possible by enabling different behaviors depending on different combinations of sensor-based conditions and sequence states. For example, just as two possible entry points into the sequence are permitted by detecting different sensor events (multi-directional or directional) at the same time as a common entry point enabling condition (homing request), branches in the sequence to different next sequenced reactive behaviors are permitted by detecting different sensor events at the same time as a common mid-sequence enabling condition (e.g., candidate behaviors each enabled by a unique combination of sensor event and sequence state). Optionally, if different sequences each are maintained with different states, different mid-sequence state combinations permit nesting of sequences (candidate behaviors each enabled by a unique combination).


Sequence planner 253 optionally maintains a navigation mode state switch or finite state machine 243a, and/or such states or requests are maintained separately (e.g., follow start, follow done).


As shown in FIG. 24B, the planner 253 may control different goal-oriented sequences 1-N, each having more than one sequenced reactive behavior therein. The behaviors therein can have as initial conditions a goal state, entry request, or other state, that the arbiter 252 identifies as requiring rendering a candidate behavior available for execution. Initial behaviors of a sequence may require a sequence request; subsequent behaviors may require either a next sequenced reactive behavior request (or, in addition, the sequence request as well).


The sequenced behaviors are “reactive” despite being dependent upon sequence state at least because they may be subject to the same arbiter 252 as the reactive behaviors, and/or react to sequence states as enabling conditions in the same manner as reactive behaviors react to real-time sensor events, and/or are prioritized lower than at least escape and avoidance behaviors. In addition, the states may be wholly reset to restart a reactive process. Even when a sequence or next sequenced reactive behavior's enabling conditions are satisfied, a reactive behavior that also has enabling conditions satisfied and preserves continuous autonomous operation will be executed instead. Reactive docking behaviors, which generally permit the robot to return and recharge for continued autonomous operation, may be prioritized higher than, in particular, navigation-related sequenced reactive behaviors. Reactive coverage behaviors may be prioritized below navigation sequenced reactive behaviors, in particular. Some portions of the docking behaviors noted herein are alternatively arranged as sequenced reactive behaviors; or other sequences are possible. Beacon-based navigation as discussed herein is an example of one implementation of combining sequenced reactive behaviors with non-sequenced reactive behaviors.


As shown in FIG. 24B, a second arbiter 252a may supervise a sequence of sequenced reactive behaviors (e.g., in combination with the sequence planner 253); or a master arbiter (not shown) may supervise a first arbiter (e.g., 252) of reactive behaviors and a subsequent arbiters (e.g., 252a) of different sequenced reactive behaviors.


As shown in FIG. 24B, for a navigation application in particular, the sequence planner 253 may include, monitor, or maintain location state in the form of a topological state map or its equivalent with respect to the environment in which the robot operates. Using beacons, this may be a topological map having nodes and links, in which the nodes may be rooms (links thereby being gateways or beacons) or gateways (links thereby being sides of the gateway). A primary use of the location state is to provide a homing beam or homing fiducial identity to the arbiter or to the sequenced reactive behaviors. If provided to the arbiter, the arbiter may recognize a sensor-based enabling condition (e.g., beam detection) for a behavior only when the beam has the correct identity (encoding) for the location state. Alternatively, the behavior itself examines the identity or encoding of a received enabling condition (e.g., beam detection). An abort condition can result if beacons have been rearranged or the like, and this abort condition can cause blacklisting of a beacon or fiducial or a reset of the map (new mapping needed). In either case, the robot 302 continues to provide coverage and all other remaining reactive behaviors in the local area. The location state is used by the planner 253 to successively execute sequences of sequenced reactive behaviors in order to traverse a distance to or from a remote or distal room (for returning from the dock for charging; for directly proceeding to a remote or distal chamber from the dock); and permits the same sequenced reactive behaviors to be used for navigation in different directions (so long as the location state is maintained) by providing the identities of the guiding fiducials or beams in the correct order for a navigation direction, depending on the direction to be traveled on the topological map.


One candidate routine maintained/executed by the planner 253 is shown in FIG. 25A. The elements of this routine are not necessarily in order (e.g., different states may be checked out of the depicted order). The planner routine monitors (502) a goal-oriented or sequence initiating condition. As discussed, this may be a variable that keeps track of a time, distance, cleanliness measure, history of obstacle interactions, or other criterion that, when satisfied, indicates that a planned sequence (such as navigation from room to room) is to be carried out. The condition may be a state, flag or state of a finite state machine. If such a condition exists (504), the planner checks (506) the state of the robot along the called for sequence, if the planner maintains such a sequence state. For example, the planner 253 may maintain or monitor the current state of a state map or finite state machine defining states of the beacon crossing sequence, as shown in FIG. 25B (in which “aligning to launch”, “homing”, “aligning after homing”, and “following field” states are available); alternatively a state map or finite state machine may be instantiated. The planner 253 may additionally or in the alternative set (508) behavior enabling conditions for entry into a sequence, such as a homing request as discussed herein, or an entire sequence entry behavior enabling condition and/or next sequenced reactive behavior enabling condition. The planner 253 may additionally set (510) guiding variables for the behavior or arbiter to recognize (e.g., such as the identity of a beam to be followed or fiducial to be homed upon or bearing to proceed in), according to, e.g., the state of a map identifying such (e.g., as shown in FIG. 25C, according to the state of a map defining which beam A or B—“green” or “red”—of lighthouses/beacons X, Y, Z connects to which rooms 1, 2, 3 and a dock, and according to the next goal room defined by the planner 253). The planner 253 may additionally monitor abort conditions (512, e.g., time out, interruption by a reactive avoidance or escape behavior) of sequenced reactive behaviors in progress, and set or reset the state of a sequence and/or the state of a map upon such an abort (although behaviors may also set or reset these internally before aborting).


After the robot 302 has detected a specified cleaning period has elapsed or the robot 302 has detected a specified number of bumps or the robot has detected a specified condition requiring a set of sequenced behaviors, the sequence planner 253 executing on the controller 250 will decide to exit the cleaning mode and enter the migration mode. The robot may also enter the migration mode right from the dock, in order to proceed to a distal or remote room in order to begin cleaning there. The migration mode executes the task of moving from one room to an adjacent room. As discussed above, the migration mode is advantageously not a monolithic process, but is a succession of sequenced reactive behaviors started by an initial behavior—that has an enabling condition of a migration mode request (also discussed herein as a “homing request”). The planner 253 will raise a “new room”/homing request enabling condition, state, or flag. The remaining enabling condition is a sensor detection of a guiding tractor beam (optionally, limited to one beam identified according to the location state 253b as originating with the lighthouse/beacon to be homed upon), which, when the beam is detected (substantially immediately if the robot is in the beam, or after the robot encounters the beam as it continues the coverage behaviors having satisfied enabling conditions), triggers the homing beam—align behavior 294a to be selected by the arbiter 252 for execution. The robot 302 performs the task of migrating from one bounded area to an adjacent bounded area by executing the homing beam—align behavior 294a while at the same time possibly executing the cliff avoiding 256 (on any occasion when any cliff detector fails to see the floor—i.e., the enabling condition of the higher priority & reactive avoidance behavior is also satisfied) and drive 266 behaviors (e.g., if an abort condition resets the sequence). When multiple behaviors are trigged by sensor inputs or the planner 253, the arbiter 252 will arbitrate the behaviors in order of priority. That is, the arbiter may also arbitrate among different sequences of reactive behaviors, or the planner may only start one sequence. The beacon crossing sequence 262 has five sequenced reactive behaviors, beginning with the homing beam—align behavior 294a, which are executed or sequentially based on controller inputs as discussed herein. If the robot 302 experiences a signal input (such as a bump sensor input) that requires execution of a behavior having a higher priority, the robot 302 will execute the behavior of higher priority and afterwards, upon the presence of the necessary enabling conditions, may reinitiate a previous sequence of reactive sequenced behaviors.



FIGS. 31A-G are overhead views showing a behavior based autonomous coverage robot 302 as it uses a navigation beacon 304 to migrate from one area to another. FIGS. 26-30 are diagrams illustrating the beacon crossing sequenced reactive behaviors 262, including beam aligning 294a, beam homing 294b, field following 294c, field aligning 294d, and leaving beacon 294e behaviors.


The beacon crossing sequence 262 commences by entering the beam homing 294b or beam aligning behaviors 294a, depending on a robot position within the first bounded area in relation to the beacon 304. Generally, the robot 302 will initiate a maneuver to detect with its omni-directional receiver 104 the first directed beam 306 from the navigational beacon 304 into a first area (e.g. turning 360 degrees or driving in a circle).



FIG. 26 depicts the beam align behavior 294a. As shown, the arbiter 252 will not permit the beam align behavior 294a as a candidate (i.e., to be prioritized with other behaviors with satisfied enabling conditions) at least until a homing request state or request is present, as well as the directional or tractor beam 306 is received by the multi or omni-directional sensor 322 of the robot 302, as behavior enabling conditions or triggers. To permit execution, no higher priority behaviors (e.g., reactive behaviors) have satisfied enabling conditions. This tractor or directed beam 306 may be limited to one expected at the present location of the robot 302 on the state map (e.g., by responding only to a beam having an identity or encoding corresponding to the current map state). Once these are satisfied, the arbiter 252 permits the beam align behavior 294a to execute. If either condition is no longer satisfied, the behavior may abort. As noted in FIG. 26, abort conditions (e.g., including missing behavior enabling conditions; time-out; or displacement by a higher priority behavior) may be monitored by the planner 253, which may also keep states reflecting the progress of the sequence as well as the location of the robot 302. The beam align behavior 294a turns or pivots the robot 302 essentially in place (522), starting in a random direction. A likely beam direction may also be maintained as a state, e.g., based on past beam detections monitored by, e.g., a concurrent process, and if so, the turning direction may be the one expected to be shorter. Once the directional receiver 318 senses the directed beam 306 (524), the behavior ends. There is no need to set a state as a request for the beam homing behavior 294b likely to follow, as the robot 302 will, so long as the homing request is active, continue to attempt to align and home when the correct beam is encountered and either is a valid entry point for the sequence. Later stages of the sequence are mid-sequence, and may use such states. Accordingly, the beam align behavior 294a is an example of choosing not to set a state in order to maintain a higher level of reactive control and likely robustness.



FIG. 31A shows the robot 302 during readiness for and execution of the beam align behavior 294a, as it encounters a first directed beam 306 emitted by a navigation beacon 304, when a homing request from the planner 293 is active. Upon detecting the first directed beam 306, the robot 302 commences the beam aligning behavior 294a. During beam aligning 294a, the drive system 130 maneuvers (e.g., by turning in place, e.g., starting in a random direction) the robot 302 to detect the first directed beam emission 306 with the directional receiver 318, which is to be aligned with a robot drive direction. By using the signals from the two component receivers 114, 116 of the directional receiver 318, the robot 302 aligns the drive direction with the directed emission 306 of the beacon 304. FIG. 31B shows the robot 302 as it detects beam 306 with the omni-directional receiver 322 and turns to face beacon 304.


After the robot 302 has aligned the drive direction with the directed emission 306 of the beacon 304 by using the directional receiver 106, the robot may commence the beam homing behavior 294b.



FIG. 27 depicts the tractor or directed beam homing behavior 294b. As shown, the arbiter 252 will not permit execution (of the highest priority candidate) until at least a homing request state or request is active, and the directional or tractor beam 306 is received by the directional sensor 318 of the robot 322 as behavior enabling conditions or triggers. This tractor or directed beam 306 may again be limited to one expected at the present location of the robot 302 on the state map. Once these are satisfied, the arbiter 252 may permit the beam homing behavior 294b to execute. If either condition is no longer satisfied, the behavior 294b may abort. As noted in FIG. 27, abort conditions and resulting states may be supervised by the planner 253. The tractor or directed beam homing behavior 294b follows the directed beam emission (532). One exemplary way of following is using alternating curves. Each parallel sensor 114 and 116 of the directional binocular sensor 318 is conditioned to return a one bit true-false value reflecting beam presence, based on signal strength (e.g., independently thresholded, or thresholded with respect to one another). As the robot 302 turns toward and away from the beam 306, beam detections of true-false, true-true, and false-true are possible (conditioning may also limit the possibilities to true-false and false-true). A “bang-bang” servoing may be used to change turning direction as the robot 302 moves forward. Each turn is an arc with decreasing radius, i.e., part of an inward spiral. Alternatively, analog values may be maintained and more complex feedback control employed. The following (532) may alternatively employ the frequency and variance tracking discussed herein with respect to docking to determine that the robot is proceeding in the correct direction and move forward in a straight line. Once the non-directional/omni-directional receiver 322 senses the directed beam 306 (534), the behavior is ready to end. A state is to be set to continue the sequence of sequenced reactive behaviors—i.e., a field aligning request. The behavior 294b may set this state itself (536), e.g., as a flag in memory; or the planner 253 may monitor that the behavior 294a is complete and set such a flag or state in a finite state map or machine.


The behavioral sequence, as discussed, permits the robot 302, after initiating the beacon crossing sequence 262, to skip the beam aligning behavior 294a if the directional receiver 318 already detects the directed beam emission 306 and is already aligned with the directed beam emission 306. Abort conditions may prevent these and permit later restarts. As discussed herein, a combination of reactive behavior based control and sequenced reactive behavior control is advantageous. However, some embodiments discussed herein do not necessarily exclude other reactive or non reactive techniques. While executing the beam homing behavior 294b, the robot 302 moves towards the navigational beacon 304 while maintaining its alignment with (e.g., servoing upon) the directed beam 306 in response to the signals of the directional receiver 318. FIG. 31C shows the robot 302 approaching the navigation beacon 304. As the robot 302 follows the edge 320 of the beam 306, the robot 302 adjusts its course as shown by arrows 308. As shown in FIG. 31C, the robot 302 stops moving towards the beacon 304 when it encounters a proximity field emission (or “force field”) 310 projected laterally from beacon 304 in all directions.


As an alternative behavior that would be suitable as part of the sequenced reactive behaviors, if the directional receiver 318 is selected to be more sensitive (e.g., by collimation and confinement) at remote distances than the omni-directional receiver 322, the directional receiver 318 may receive the force field emission 310 even though the omni-directional sensor 322 does not. In this case, the directional sensor 318 may be used to alternatively home on the force field emission 310. In this case, an alternative behavior is provided in the sequence, identical to the tractor homing behavior 294b except that the sensor-based enabling condition is that the force field 310 is seen in the directional sensor 318; and the homing is on this emission 310. As an example of behavioral robustness, this behavior may be arbitrarily added to the sequence of reactive behaviors without changing any other behavior or disrupting any other process, with a priority above or below the existing tractor homing behavior 294a depending on empirical evidence that one or the other tended to be more reliable.


Upon detecting the proximity field 310, the robot 302 executes the field aligning behavior 294c. FIG. 28 depicts the field align behavior 294c. As shown, the arbiter 252 will not permit execution (of the highest priority candidate) until at least a homing request state or request is active, and either a field align request is active, or the proximity field 310 is seen in the non-directional sensor 322 (this latter permitting the behavior sequence to advance to field aligning 294c when the robot 302 fortuitously “stumbles” into the proximity field 310 before encountering the directed beam 306). The field may again be limited to one expected at the present location of the robot 302 on the state map. Once these are satisfied, the arbiter 252 may permit the field align behavior 294c to execute, abort conditions supervised as previously noted by the planner 253. The field align behavior 294c turns or pivots the robot essentially in place (544). A likely beam direction may be determined according to the beam identity (e.g., facing the dock 340, following a left beam 350 to the dock 340 likely means the robot 302 should turn clockwise, following a right beam 360 counterclockwise) or other state or information. The robot 302 may turn in the direction expected to be shorter or randomly. Once the omni or non-directional receiver 322 senses the directed beam 306 (546), the behavior 294c ends. Optionally, the behavior 294c does not end until an additional condition is satisfied—that the directional beam 306 is not present in the directional receiver 318 (which increases the likelihood that the robot 302 is rotated transversely to the dock 340). A field follow request is to be set to continue the sequence of sequenced reactive behaviors, as discussed, by the behavior itself (548), by the planner 253, or otherwise.


Accordingly, as shown in FIG. 31D, the robot 302 can detect the proximity field 310 with the directional receiver 318, as when the robot 302 is facing the beacon 304, the robot 302 will initiate a pivot maneuver (e.g., turn in place for differential drive robots having drive wheels on the center diameter or holonomic drives or highly steerable robots, turn substantially in place for other configurations) altering the robot drive direction to the right or left toward the other directed beam 314, based on the beam encoding of the first directed beam 306. The robot 302 stops the pivot maneuver when it no longer detects the proximity beam 310 with the directional receiver 106.


The robot 302 may encounter the proximity field 310 before encountering the directed beam 306. In this case, the robot 302 does not execute the beam aligning or beam homing behavior, 294a and 294b, respectively. Instead, the robot 302 recognizes the encoding of the proximity field 310 and executes the field aligning behavior 294d.


When the robot 302 no longer detects the proximity field beam 310 with the directional receiver 218, the robot 302 executes the field following behavior 294c.



FIG. 29 depicts the field following behavior 294d. As shown, the arbiter 252 will not permit execution (of the highest priority candidate) until at least a homing request state or request is active, and a field follow request is active. As noted, this may be limited to simply the field follow request if the sequence is to be more readily abandoned to preserve robustness. The field is optionally limited to an expected one. Once these are satisfied, the arbiter 252 may permit the field follow behavior 294d to execute subject to abort conditions. The field follow behavior 294c follows the edge 312 of the proximity field emission 310 (554) using alternating, decreasing radius curves as previously discussed. The non-directional sensor 322 may be conditioned to return a one bit true-false value reflecting beam presence, based on signal strength. Once the non-directional/omni-directional receiver 322 senses the next directed beam 314 (534), i.e., the launch beam (this detection limited to the expected beam of the beacon 304 followed and/or according to the location state of the robot 304), the behavior is ready to end. Optionally, the behavior 294d further checks and proceeds until the directional receiver 318 does not sense the launch beam 314 (which tends to exclude those cases in which the omni-directional detector 322 detects a launch beam 314 too early via a reflection from a highly reflective surface, such as a nearby white door or wall corner). A state is to be set to continue the sequence of sequenced reactive behaviors—i.e., a launch request. The behavior 294c may set this state itself (558), or the planner 253 may control this.


As shown in FIG. 31E, during the field following behavior 294d, the robot 302 maneuvers in an resultant arc around the beacon 304 while following an edge 312 of the proximity beam 310 via omni-directional receiver 322. The robot 302 follows the edge 312 of the force field 310 by monitoring the received signal strength of the force field emission 310 on the robot's emission sensors 318 and 322. One example of the arc maneuver entails the robot 302 “zigzagging” back and forth (e.g., “bang-bang” servoing or following using a one-bit true-false detection and forward motion in initially arced, steadily decreasing turn radius, e.g., spiraling in) along the edge 312 of the proximity beam 310, where the robot 302 detects entering and exiting the field of emission of the proximity beam 310 with the omni-directional receiver 322. In another example, the robot 302 follows the edge 312 of the force field by using a PID control system based on the time in the force field 310 or the time out of the force field 310. The PID control system controls the arc of the robot 302. The robot 302 executes the maneuver at a reduced speed (e.g., less than 50% of coverage speed) for both drive wheels 132 and 134, so that both drive wheels 132 and 134 are being driven in a forward direction. This facilitates easier traversing of transitions and thresholds. The robot 302 passes through the area normally covered by a gateway beam 316. The robot 302 continues to follow the edge 312 of the proximity beam 310 until the omni-directional receiver 322 detects that the robot 302 has entered the second directional beam 314 and the directional receiver 318 no longer detects the second directed beam 314, as shown in FIG. 31E. As noted, satisfying both criteria helps prevent the robot 302 from detecting reflected beam emissions off of adjacent objects, such as a white door.



FIG. 29 depicts the launch behavior 294e. As shown, the arbiter 252 will not permit execution (of the highest priority candidate) until the launch request is active, and may also require the homing request. The launch behavior 294e assumes a last position with the omni-directional detector 322 in the launch beam 314 and a bearing substantially transverse to the launch beam 314, and curves away from the field 314/beacon 304 in the approximate direction of the launch beam 314 (e.g., facing the dock 340, if the robot 302 has traversed left to right, curving clockwise, and if the robot 202 has traversed the opposite direction, curving counterclockwise). Optionally, the robot 302 may try to stay within the launch beam 314 by servoing to return to the launch beam 314 when the launch beam 314 is not detected in the omni-directional sensor 322. After a predetermined time (e.g., 5 s) or distance (e.g., 1 m) or obstacle encounter, the behavior is ready to end. At this point, the planner 253 may terminate the homing request (and, e.g., updating the location state of the topological map as well as clearing or resetting a navigation mode state if one is kept), and the reactive behaviors will begin the robot's coverage work in the subsequent chamber (or, if the battery is to be charged, this may initiate another homing request in the direction of the dock according to the location state of the map). The planner 253 may also immediately initiate another homing request; or reset variables of the current homing request according to the location state of the map such that the robot 302 will proceed from room to room to room, e.g., moving directly to or from a distal or remote room.


That is, as shown in FIG. 31F, upon completing the field following behavior 294c, the robot 302 may execute the leaving beacon behavior 294e, as shown in FIG. 30, which entails the robot 302 moving away from the beacon 304. Alternatively, this can be performed as a pivot and launch, and/or guided by the launch beam 314 and omni-directional receiver 322, e.g., FIG. 31G shows the robot 302 turning away from navigation beacon 304 to face a second area as the robot 302 passes the edge of directed beam 314. If the robot 302 detects the second directed beam emission 314 with the omni-directional receiver 322, the robot 302 maneuvers away from the beacon 304. If the robot 302 does not detect the second directed beam emission 314 with the omni-directional receiver 322, the robot 302 moves in a forward direction as shown in FIG. 31H without turning in order to avoid arcing back into the gateway. Another example of the leave beacon behavior 294e includes the robot 302 aligning itself in a particular direction using odometry (or a next beacon in a sequence, e.g., when the planner 253 includes a distal or remote room traversal sequence that restarts the sequence of behaviors at beam align with a next beacon's beam or field according to the map) and moving in that direction without guidance from any of the beams emitted from the current beacon 304. FIG. 31G shows the robot 302 moving away from navigation beacon 304 as it finishes migrating.


While migrating, the robot 302 stores in memory the beam encodings of the first and second directed beams 306 and 314, respectively, at each navigational beacon 304 encountered while away from a base station 340. To return to the base station 340, the robot 302 conducts coverage behaviors while the homing request is on, resulting in discovering the appropriate directed beams 306 and 314, respectively, and corresponding encodings, to follow a sequence of encountered beams back to the base station 340. In one example, the robot 302 can wirelessly communicate (e.g. radiofrequency) with the navigational beacons 304 to independently activate each previously encountered beam 306, 314, and 316 one at a time to find (traversing a room using random bounce and/or obstacle following to place itself in a beam path) and follow each beam back to the base station 340.


In another example (not shown), navigation beacons may emit at least one, two, or three beam signals: fence/gate beam (referred to herein as “yellow”, but in reality having a first modulation of a first frequency in the IR band), right-hand navigation beam (referred to as “red”, but in reality having either or both of different modulation or a different frequency in the IR band from the “yellow” beam), and left-hand navigation beam (referred to as “green”, but in reality having either or both of a different modulation or a different frequency in the IR band from the “yellow” and “red” beams).


If multiple navigation beacons are provided, each would have different modulation of its beams, but the emitters of the fence/gate beam, right hand beam, or left hand beam would preferably be consistent among the navigation beacons.


Regardless of its assignment as fence, gate, or trail marker, as one example, a navigation beacon also may emit a geometrically determined force field (“blue”, again in actuality either or both of a different modulation or frequency in the IR band, although a force field could also be radio frequency RF rather than visible or IR light) in a circular region around their base. This may identify its presence to a robot, and allow a robot to follow a path around the Navigation beacon (through a gate or past a trail marker), for example.


An IR-based beacon system can be designed to use a geometric solution, similar in certain respects to a non-limiting example such as a shaded floor lamp which lights up only a small area around itself, in order to provide a well-defined force field via IR, for example.


A fence navigation beacon may emit a yellow beam directly out the front of the navigation beacon (0 degrees) and a single navigation beam (red or green) at about a 60-degree angle to the right or left of the yellow beam, for example. A gate navigation beacon may emit a yellow beam to the front, a red beam at −60 degrees, and a green beam at about +60 degrees, as another non-limiting example; also, for example, a trail marker Navigation beacon may emit a red beam at about −60 degrees and a green beam at about +60 degrees.


Navigation beacon “beams” in some non-limiting examples may not be narrowly focused (thus not requiring a special lens); they may fan out to light as much of the room as may be practicable without mutual interference. As an alternative, for example, a standard home base LED not requiring other optics may also be used.


Referring to FIGS. 24A-B and 32A-E, the docking behavior priorities (from highest to lowest) include docking retry docking bump follow 264c, docking bounce 264d, docking quick turn 296c, docking distant homing 296b, docking lobe following 296a, and docking field following 296d. The robot 302 must generally approach the dock 340 from a forward direction with less than 4 degrees of skew in order to dock properly.


The default docking behavior, docking lobe following 296a, as shown in FIG. 32A, is executed when no other docking behavior has a higher priority. When the robot 302 detects the left (red) beam 350 only, it follows the edge of the red beam 350 in a clockwise direction. When the robot 302 detects the right (green) beam 360 or the overlap area of the red and green beams 350 and 360, respectively, it follows the edge of the green beam 360 in a counterclockwise direction. This results in the robot 302 appearing to follow the outside of the nearest docking beam 350 or 360 around to the front of the dock 340 at zero degrees and then following a normal (zero degree) edge 362 right onto the docking contacts. When the robot 302 detects the force field 370 during this behavior it slows down in order to follow more accurately.


The robot 302 keeps track of the signals detected by the omni-directional receiver 322 over a window of 30 seconds. When the robot 302 detects that the frequency and variance of line crossing events (crossing the normal (zero degree) edge 362) is above about 1 Hz and below a variance of about 2.5 seconds it determines that the robot 302 is following the normal (zero degree) edge 362 and will simply drive straight by executing a smooth homing behavior, instead of continuing to follow the edge 362 (with the inevitable right-left oscillation) under the lobe following behavior 296a.


Referring to FIG. 32B, the quick turn behavior 296c is activated when the robot 302 has encountered a docking beam 350 or 360 while driving straight, spiraling, bump following, or wall following, is not in the force field 370, and has not detected the dock 340 with its directional receivers 318 in the last several seconds. When the robot 302 detects the red beam 350, it arcs clockwise toward the dock 340. When the robot 302 detects the green beam 360 or the overlap area of the red and green beams 350 and 360, it arcs counterclockwise toward the dock 340. When the robot 302 detects the other beam 350 or 360, detects the dock 340 with its directional receivers 318, or has arced more than 360 degrees, this behavior 296c aborts. Typically it is then followed by the lobe following behavior 296a.


Referring to FIG. 32C, the distant homing behavior 296b is activated when the robot 302 has not detected the force field 370 in its omni-directional receiver 322 in the last several seconds, and detects the dock 340 with the directional receiver 318. If the robot 340 detects the red or green beams 350 and 360, respectively, it will drive toward them. If the robot 340 only detects the force field 370 it will drive toward the force field 370. This allows the robot 302 to approach the dock 340 from a distance at any angle. When the robot 302 detects the force field 370 with its omni-directional receiver 322 this behavior 296b aborts. During a front approach this is typically followed by the lobe following behavior 296a. During a side approach this is typically followed by the field following behavior 296d.


Referring to FIG. 32D, the field following behavior 296b activates when the robot 302 detects the force field 370 with its omni-directional receiver 322, has not detected the red or green beam, 350 and 360 respectively, with its omni-directional receiver 322 in the last second, and does not see the red or green beam 350 and 360, respectively, with its directional receivers 318. The robot 302 follows the edge 372 of the force field 370 with its omni-directional receiver 322 in a random direction (cw or ccw) until it detects the red or green beam, 350 and 360 respectively, a bump, a cliff, or has traveled more than 1 meter. When the robot 302 detects the red or green beam, 350 and 360 respectively, it engages in reverse lobe following 296a for a distance of about 1 meter or until a bump is detected in order to travel away from the dock 340 along the normal (zero degree) edge 362 and straighten out. Then the robot 302 turns 180 degrees or until it faces the dock 340. Then the behavior 296b aborts. Typically the lobe following behavior 296a activates next to finish docking.


Referring to FIG. 32E, the robot 302 can dock successfully even when there are obstacles completely or partially occluding the docking beams 350 and 360 or the robot's physical access to the dock 340 from some, but not all, approach angles by executing an obstacle dock behavior 296e. When the robot 302 is seeking the dock 340 and has detected a docking beam 350 or 360 in its omni-directional receiver 322 in the last several seconds, and then also detects a bump, it activates the docking bounce behavior 264c. With about a 66% probability, the obstacle dock behavior 296e initiates a bump follow behavior 264c and with about a 33% probability, the docking bounce behavior 264d which has the robot 302 simply back up, turn a random angle between 10 and 170 degrees for side impacts or 75 and 255 degrees for front impacts, and then aborts the behavior 264d. The bump follow behavior 264c follows an edge of the obstacle with the robot's bumper until one of the following conditions occurs: a) the robot 302 detects the dock 340 with the directional receiver 318 and does not detect it with is omni-directional receiver 322, with about a 66% probability; b) the robot 302 detects a crossing of the normal (zero degree) edge 362 from right to left and is bump following 264c in a clockwise direction, or the robot 302 detects a crossing of the normal (zero degree) edge 362 from left to right and is bump following 264c in a counterclockwise direction, with about a 66% probability; c) the robot 362 detects the force field 370 with its omni-directional receiver 322; d) 30 seconds has passed since the onset of the bump follow behavior 264c; or e) no docking beam 350 or 360 has been detected by the omni-directional receiver 322 for more than about 5 seconds.


Another possible implementation of bump follow 264c includes the robot 302 keeping track of the projected location of the dock 340 and turning preferentially toward the estimated dock location during bump follow 264c. Each time the robot 302 crosses the normal (zero degree) edge 362 while facing the dock 340 and detecting the dock 340 by its directional receiver 318, the robot 302 uses odometry to project the location of the dock 340 about 5 feet out along a direction of travel. The robot 302 can use odometry throughout the docking maneuver to estimate the orientation of itself to the projected location of the dock 340.


The combination of short bouts of bump following and normal docking methods allow the robot 302 to dock in the face of a large variety of obstacles including but not limited to walls, chairs, and boxes, as well as reflections of the IR docking beams.


“ROBOT OBSTACLE DETECTION SYSTEM”, U.S. Pat. No. 6,594,844, disclosing proximity sensors such as cliff sensors and wall following sensors; “AUTONOMOUS FLOOR-CLEANING ROBOT”, U.S. Pat. No. 6,883,201, disclosing a general structure of an iRobot Roomba coverage/cleaning robot and main and edge cleaning heads in detail; “METHOD AND SYSTEM FOR MULTI-MODE COVERAGE FOR AN AUTONOMOUS ROBOT”, U.S. Pat. No. 6,809,490, disclosing motion control and coverage behaviors, including escape behaviors, selected by an arbiter according to the principles of behavior based robotics; and “METHOD AND SYSTEM FOR ROBOT LOCALIZATION AND CONFINEMENT”, U.S. Pat. No. 6,781,338, disclosing virtual walls, i.e., robot confinement using wall-simulating directed beams, are each incorporated by reference herein in their entireties.


Other robot details and features combinable with those described herein may be found in the following U.S. patent applications filed concurrently herewith, entitled “COVERAGE ROBOT MOBILITY” having assigned Ser. No. 11/633,885; “MODULAR ROBOT” having assigned Ser. No. 11/633,886; and “ROBOT SYSTEM” having assigned Ser. No. 11/633,883, the entire contents of the aforementioned applications are hereby incorporated by reference.


A number of implementations have been described. Nevertheless, it will be understood that various modifications may be made without departing from the spirit and scope of the following claims. Accordingly, other implementations are within the scope of the following claims.

Claims
  • 1. A cleaning system for cleaning a room, the cleaning system comprising: an autonomous mobile robot comprising: a chassis;drive wheels connected to the chassis for moving the mobile robot to a docked position;a central receiver and two side receivers located at a front portion of the chassis; anda controller located within the chassis and configured to receive signals from at least two of the receivers and direct the mobile robot toward a docked position using the receiver signals; anda base station comprising three infrared emitters, the three infrared emitters comprising at least two side emitters spaced in opposite directions from a central location of the base station, at least two of the emitters being detectable by the mobile robot from a position proximate the base station when the mobile robot is aligning itself for docking,wherein the receivers of the mobile robot are configured to receive infrared directed beams from at least the two side emitters of the base station, and the controller of the mobile robot is configured to direct the mobile robot toward a docked position on the based station using the infrared directed beams received from the two side emitters when the robot is in a docking mode, the controller including a control routine for docking the mobile robot from a position within the room wherein the mobile robot has been aligned for docking but is outside of a range of an infrared directed beam that maintains contact with the mobile robot until the mobile robot is docked.
  • 2. The system of claim 1, wherein the controller can differentiate one infrared directed beam from the other infrared directed beam.
  • 3. The system of claim 2, wherein the beams are differentiated by time division multiplexing with a fixed period random offset.
  • 4. The system of claim 2, wherein the beams are differentiated by encoding of the beams.
  • 5. The system of claim 4, wherein the beams are encoded by modulation with differing serial codes.
  • 6. The system of claim 2, wherein the side emitters are arranged on the base station to emit laterally bounded and overlapping fields of emissions of signal beams.
  • 7. The system of claim 2, wherein each of the side emitters projects a beam of about twelve degrees from a direction normal to the emitter.
  • 8. The system of claim 2, wherein the side emitters are positioned such that their angle of overlap is between about six degrees and about ten degrees, and each emitter's aperture angle is about twenty to about thirty degrees.
  • 9. A method for navigating a mobile cleaning robot to a docked position on a base station having three emitters, at least two of the base station emitters being side emitters spaced in opposite directions from a central location of the base station, the method comprising: receiving infrared directed beams from the side emitters when the robot is in a docking mode and until the robot is docked, the infrared directed beams being indicative of a docked position; anddriving the mobile robot to the docked position on the base station using the infrared directed beams by aligning the robot for docking when the robot is within range of all three emitters.
  • 10. The method of claim 9, further comprising charging a power source of the mobile robot when the mobile cleaning robot is in a docked position.
  • 11. The method of claim 9, further comprising differentiating one infrared directed beam from the other infrared directed beam.
  • 12. The method of claim 11, further comprising differentiating the infrared directed beams by time division multiplexing with a fixed period random offset.
  • 13. The method of claim 11, further comprising differentiating the infrared directed beams by encoding the beams.
  • 14. The method of claim 9, further comprising identifying each infrared directed beam by its modulation.
  • 15. The method of claim 9, wherein receiving infrared directed beams from the side emitters comprises receiving infrared directed beams within a sweep angle of between about twenty eight degrees and about thirty three degrees.
Parent Case Info

This is a continuation of U.S. Non-Provisional Patent Application Ser. No. 11/633,869, filed Dec. 4, 2006, which claims priority to U.S. Provisional Patent Application No. 60/741,442, filed Dec. 2, 2005, the disclosures of which are incorporated herein by reference in their entireties.

US Referenced Citations (997)
Number Name Date Kind
1755054 Darst Apr 1930 A
1780221 Buchmann Nov 1930 A
1970302 Gerhardt Aug 1934 A
2136324 John Nov 1938 A
2302111 Dow et al. Nov 1942 A
2353621 Sav et al. Jul 1944 A
2770825 Pullen Nov 1956 A
2930055 Fallen et al. Mar 1960 A
3119369 Harland et al. Jan 1964 A
3166138 Dunn Jan 1965 A
3333564 Waters Aug 1967 A
3375375 Robert et al. Mar 1968 A
3381652 Schaefer et al. May 1968 A
3457575 Bienek Jul 1969 A
3550714 Bellinger Dec 1970 A
3569727 Aggarwal et al. Mar 1971 A
3649981 Woodworth Mar 1972 A
3674316 De Brey Jul 1972 A
3678882 Kinsella Jul 1972 A
3690559 Rudloff Sep 1972 A
3744586 Leinauer Jul 1973 A
3756667 Bombardier et al. Sep 1973 A
3809004 Leonheart May 1974 A
3816004 Bignardi Jun 1974 A
3845831 James Nov 1974 A
3851349 Lowder Dec 1974 A
3853086 Asplund Dec 1974 A
3863285 Hukuba Feb 1975 A
3888181 Kups Jun 1975 A
3937174 Haaga Feb 1976 A
3952361 Wilkins Apr 1976 A
3989311 Debrey Nov 1976 A
3989931 Phillips Nov 1976 A
4004313 Capra Jan 1977 A
4012681 Finger et al. Mar 1977 A
4070170 Leinfelt Jan 1978 A
4099284 Shinozaki et al. Jul 1978 A
4119900 Kremnitz Oct 1978 A
4175589 Nakamura et al. Nov 1979 A
4175892 De Nov 1979 A
4196727 Verkaart et al. Apr 1980 A
4198727 Farmer Apr 1980 A
4199838 Simonsson Apr 1980 A
4209254 Reymond et al. Jun 1980 A
D258901 Keyworth Apr 1981 S
4297578 Carter Oct 1981 A
4305234 Pichelman Dec 1981 A
4306329 Yokoi Dec 1981 A
4309758 Halsall et al. Jan 1982 A
4328545 Halsall et al. May 1982 A
4367403 Miller Jan 1983 A
4369543 Chen et al. Jan 1983 A
4401909 Gorsek Aug 1983 A
4416033 Specht Nov 1983 A
4445245 Lu May 1984 A
4465370 Yuasa et al. Aug 1984 A
4477998 You Oct 1984 A
4481692 Kurz Nov 1984 A
4482960 Pryor Nov 1984 A
4492058 Goldfarb et al. Jan 1985 A
4513469 Godfrey et al. Apr 1985 A
D278732 Ohkado May 1985 S
4518437 Sommer May 1985 A
4534637 Suzuki et al. Aug 1985 A
4556313 Miller et al. Dec 1985 A
4575211 Matsumura et al. Mar 1986 A
4580311 Kurz Apr 1986 A
4601082 Kurz Jul 1986 A
4618213 Chen Oct 1986 A
4620285 Perdue Oct 1986 A
4624026 Olson et al. Nov 1986 A
4626995 Lofgren et al. Dec 1986 A
4628454 Ito Dec 1986 A
4638445 Mattaboni Jan 1987 A
4644156 Takahashi et al. Feb 1987 A
4649504 Krouglicof et al. Mar 1987 A
4652917 Miller Mar 1987 A
4654492 Koerner et al. Mar 1987 A
4654924 Getz et al. Apr 1987 A
4660969 Sorimachi et al. Apr 1987 A
4662854 Fang May 1987 A
4674048 Okumura Jun 1987 A
4679152 Perdue Jul 1987 A
4680827 Hummel Jul 1987 A
4696074 Cavalli et al. Sep 1987 A
D292223 Trumbull Oct 1987 S
4700301 Dyke Oct 1987 A
4700427 Knepper Oct 1987 A
4703820 Reinaud Nov 1987 A
4709773 Clement et al. Dec 1987 A
4710020 Maddox et al. Dec 1987 A
4712740 Duncan et al. Dec 1987 A
4716621 Zoni Jan 1988 A
4728801 O'Connor Mar 1988 A
4733343 Yoneda et al. Mar 1988 A
4733430 Westergren Mar 1988 A
4733431 Martin Mar 1988 A
4735136 Lee et al. Apr 1988 A
4735138 Gawler et al. Apr 1988 A
4748336 Fujie et al. May 1988 A
4748833 Nagasawa Jun 1988 A
4756049 Uehara Jul 1988 A
4767213 Hummel Aug 1988 A
4769700 Pryor Sep 1988 A
4777416 George, II et al. Oct 1988 A
D298766 Tanno et al. Nov 1988 S
4782550 Jacobs Nov 1988 A
4796198 Boultinghouse et al. Jan 1989 A
4806751 Abe et al. Feb 1989 A
4811228 Hyyppa Mar 1989 A
4813906 Matsuyama et al. Mar 1989 A
4815157 Tsuchiya Mar 1989 A
4817000 Eberhardt Mar 1989 A
4818875 Weiner Apr 1989 A
4829442 Kadonoff et al. May 1989 A
4829626 Harkonen et al. May 1989 A
4832098 Palinkas et al. May 1989 A
4851661 Everett Jul 1989 A
4854000 Takimoto Aug 1989 A
4854006 Nishimura et al. Aug 1989 A
4855915 Dallaire Aug 1989 A
4857912 Everett et al. Aug 1989 A
4858132 Holmquist Aug 1989 A
4867570 Sorimachi et al. Sep 1989 A
4880474 Koharagi et al. Nov 1989 A
4887415 Martin Dec 1989 A
4891762 Chotiros Jan 1990 A
4893025 Lee Jan 1990 A
4901394 Nakamura et al. Feb 1990 A
4905151 Weiman et al. Feb 1990 A
4909972 Britz Mar 1990 A
4912643 Beirne Mar 1990 A
4918441 Bohman Apr 1990 A
4919224 Shyu et al. Apr 1990 A
4919489 Kopsco Apr 1990 A
4920060 Parrent et al. Apr 1990 A
4920605 Takashima May 1990 A
4933864 Evans et al. Jun 1990 A
4937912 Kurz Jul 1990 A
4953253 Fukuda et al. Sep 1990 A
4954962 Evans et al. Sep 1990 A
4955714 Stotler et al. Sep 1990 A
4956891 Wulff Sep 1990 A
4961303 McCarty et al. Oct 1990 A
4961304 Ovsborn et al. Oct 1990 A
4962453 Pong et al. Oct 1990 A
4967862 Pong et al. Nov 1990 A
4971591 Raviv et al. Nov 1990 A
4973912 Kaminski et al. Nov 1990 A
4974283 Holsten et al. Dec 1990 A
4977618 Allen Dec 1990 A
4977639 Takahashi et al. Dec 1990 A
4986663 Cecchi et al. Jan 1991 A
5001635 Yasutomi et al. Mar 1991 A
5002145 Waqkaumi et al. Mar 1991 A
5012886 Jonas et al. May 1991 A
5018240 Holman May 1991 A
5020186 Lessig et al. Jun 1991 A
5022812 Coughlan et al. Jun 1991 A
5023788 Kitazume et al. Jun 1991 A
5024529 Svetkoff et al. Jun 1991 A
D318500 Malewicki et al. Jul 1991 S
5032775 Mizuno et al. Jul 1991 A
5033151 Kraft et al. Jul 1991 A
5033291 Podoloff et al. Jul 1991 A
5040116 Evans et al. Aug 1991 A
5045769 Everett Sep 1991 A
5049802 Mintus et al. Sep 1991 A
5051906 Evans et al. Sep 1991 A
5062819 Mallory Nov 1991 A
5070567 Holland Dec 1991 A
5084934 Lessig et al. Feb 1992 A
5086535 Grossmeyer et al. Feb 1992 A
5090321 Abouav Feb 1992 A
5093955 Blehert et al. Mar 1992 A
5094311 Akeel Mar 1992 A
5098262 Wecker et al. Mar 1992 A
5105502 Takashima Apr 1992 A
5105550 Shenoha Apr 1992 A
5109566 Kobayashi et al. May 1992 A
5111401 Everett et al. May 1992 A
5115538 Cochran et al. May 1992 A
5127128 Lee Jul 1992 A
5136675 Hodson Aug 1992 A
5136750 Takashima et al. Aug 1992 A
5142985 Stearns et al. Sep 1992 A
5144471 Takanashi et al. Sep 1992 A
5144714 Mori et al. Sep 1992 A
5144715 Matsuyo et al. Sep 1992 A
5152028 Hirano Oct 1992 A
5152202 Strauss Oct 1992 A
5155684 Burke et al. Oct 1992 A
5163202 Kawakami et al. Nov 1992 A
5163320 Goshima et al. Nov 1992 A
5164579 Pryor et al. Nov 1992 A
5165064 Mattaboni Nov 1992 A
5170352 McTamaney et al. Dec 1992 A
5173881 Sindle Dec 1992 A
5182833 Yamaguchi et al. Feb 1993 A
5187662 Kamimura et al. Feb 1993 A
5202742 Frank et al. Apr 1993 A
5204814 Noonan et al. Apr 1993 A
5206500 Decker et al. Apr 1993 A
5208521 Aoyama May 1993 A
5216777 Moro et al. Jun 1993 A
5222786 Sovis et al. Jun 1993 A
5227985 DeMenthon Jul 1993 A
5233682 Abe et al. Aug 1993 A
5239720 Wood et al. Aug 1993 A
5251358 Moro et al. Oct 1993 A
5261139 Lewis Nov 1993 A
5276618 Everett Jan 1994 A
5276939 Uenishi Jan 1994 A
5277064 Knigga et al. Jan 1994 A
5279672 Betker et al. Jan 1994 A
5284452 Corona Feb 1994 A
5284522 Kobayashi et al. Feb 1994 A
5293955 Lee Mar 1994 A
D345707 Alister Apr 1994 S
5303448 Hennessey et al. Apr 1994 A
5307273 Oh et al. Apr 1994 A
5309592 Hiratsuka May 1994 A
5310379 Hippely et al. May 1994 A
5315227 Pierson et al. May 1994 A
5319827 Yang Jun 1994 A
5319828 Waldhauser et al. Jun 1994 A
5321614 Ashworth Jun 1994 A
5323483 Baeg Jun 1994 A
5324948 Dudar et al. Jun 1994 A
5331713 Tipton Jul 1994 A
5341186 Kato Aug 1994 A
5341540 Soupert et al. Aug 1994 A
5341549 Wirtz et al. Aug 1994 A
5345649 Whitlow Sep 1994 A
5352901 Poorman Oct 1994 A
5353224 Lee et al. Oct 1994 A
5363305 Cox et al. Nov 1994 A
5363935 Schempf et al. Nov 1994 A
5369347 Yoo Nov 1994 A
5369838 Wood et al. Dec 1994 A
5386862 Glover et al. Feb 1995 A
5399951 Lavallee et al. Mar 1995 A
5400244 Watanabe et al. Mar 1995 A
5404612 Ishikawa Apr 1995 A
5410479 Coker Apr 1995 A
5435405 Schempf et al. Jul 1995 A
5440216 Kim Aug 1995 A
5442358 Keeler et al. Aug 1995 A
5444965 Colens Aug 1995 A
5446356 Kim Aug 1995 A
5446445 Bloomfield et al. Aug 1995 A
5451135 Schempf et al. Sep 1995 A
5454129 Kell Oct 1995 A
5455982 Armstrong et al. Oct 1995 A
5465525 Mifune et al. Nov 1995 A
5465619 Sotack et al. Nov 1995 A
5467273 Faibish et al. Nov 1995 A
5471560 Allard et al. Nov 1995 A
5491670 Weber Feb 1996 A
5497529 Boesi Mar 1996 A
5498948 Bruni et al. Mar 1996 A
5502638 Takenaka Mar 1996 A
5505072 Oreper Apr 1996 A
5507067 Hoekstra et al. Apr 1996 A
5510893 Suzuki Apr 1996 A
5511147 Abdel Apr 1996 A
5515572 Hoekstra et al. May 1996 A
5534762 Kim Jul 1996 A
5535476 Kresse et al. Jul 1996 A
5537017 Feiten et al. Jul 1996 A
5537711 Tseng Jul 1996 A
5539953 Kurz Jul 1996 A
5542146 Hoekstra et al. Aug 1996 A
5542148 Young Aug 1996 A
5546631 Chambon Aug 1996 A
5548511 Bancroft Aug 1996 A
5551119 Worwag Sep 1996 A
5551525 Pack et al. Sep 1996 A
5553349 Kilstrom et al. Sep 1996 A
5555587 Guha Sep 1996 A
5560077 Crotchett Oct 1996 A
5568589 Hwang Oct 1996 A
D375592 Ljunggren Nov 1996 S
5608306 Rybeck et al. Mar 1997 A
5608894 Kawakami et al. Mar 1997 A
5608944 Gordon Mar 1997 A
5610488 Miyazawa Mar 1997 A
5611106 Wulff Mar 1997 A
5611108 Knowlton et al. Mar 1997 A
5613261 Kawakami et al. Mar 1997 A
5613269 Miwa Mar 1997 A
5621291 Lee Apr 1997 A
5622236 Azumi et al. Apr 1997 A
5634237 Paranjpe Jun 1997 A
5634239 Tuvin et al. Jun 1997 A
5636402 Kubo et al. Jun 1997 A
5642299 Hardin et al. Jun 1997 A
5646494 Han Jul 1997 A
5647554 Ikegami et al. Jul 1997 A
5650702 Azumi Jul 1997 A
5652489 Kawakami Jul 1997 A
5682313 Edlund et al. Oct 1997 A
5682839 Grimsley et al. Nov 1997 A
5696675 Nakamura et al. Dec 1997 A
5698861 Oh Dec 1997 A
5709007 Chiang Jan 1998 A
5710506 Broell et al. Jan 1998 A
5714119 Kawagoe et al. Feb 1998 A
5717169 Liang et al. Feb 1998 A
5717484 Hamaguchi et al. Feb 1998 A
5720077 Nakamura et al. Feb 1998 A
5732401 Conway Mar 1998 A
5735017 Barnes et al. Apr 1998 A
5735959 Kubo et al. Apr 1998 A
5742975 Knowlton et al. Apr 1998 A
5745235 Vercammen et al. Apr 1998 A
5752871 Tsuzuki May 1998 A
5756904 Oreper et al. May 1998 A
5761762 Kubo et al. Jun 1998 A
5764888 Bolan et al. Jun 1998 A
5767437 Rogers Jun 1998 A
5767960 Orman Jun 1998 A
5777596 Herbert Jul 1998 A
5778486 Kim Jul 1998 A
5781697 Jeong Jul 1998 A
5781960 Kilstrom et al. Jul 1998 A
5784755 Karr et al. Jul 1998 A
5786602 Pryor et al. Jul 1998 A
5787545 Colens Aug 1998 A
5793900 Nourbakhsh et al. Aug 1998 A
5794297 Muta Aug 1998 A
5802665 Knowlton et al. Sep 1998 A
5812267 Everett, Jr. et al. Sep 1998 A
5814808 Takada et al. Sep 1998 A
5815880 Nakanishi Oct 1998 A
5815884 Imamura et al. Oct 1998 A
5819008 Asama et al. Oct 1998 A
5819360 Fujii Oct 1998 A
5819936 Saveliev et al. Oct 1998 A
5820821 Kawagoe et al. Oct 1998 A
5821730 Drapkin Oct 1998 A
5825981 Matsuda Oct 1998 A
5828770 Leis et al. Oct 1998 A
5831597 West et al. Nov 1998 A
5836045 Anthony et al. Nov 1998 A
5839156 Park et al. Nov 1998 A
5839532 Yoshiji et al. Nov 1998 A
5841259 Kim et al. Nov 1998 A
5867800 Leif Feb 1999 A
5867861 Kasen et al. Feb 1999 A
5869910 Colens Feb 1999 A
5894621 Kubo Apr 1999 A
5896611 Haaga Apr 1999 A
5903124 Kawakami May 1999 A
5905209 Oreper May 1999 A
5907886 Buscher Jun 1999 A
5910700 Crotzer Jun 1999 A
5911260 Suzuki Jun 1999 A
5916008 Wong Jun 1999 A
5924167 Wright et al. Jul 1999 A
5926909 McGee Jul 1999 A
5933102 Miller et al. Aug 1999 A
5933913 Wright et al. Aug 1999 A
5935179 Kleiner et al. Aug 1999 A
5935333 Davis Aug 1999 A
5940346 Sadowsky et al. Aug 1999 A
5940927 Haegermarck et al. Aug 1999 A
5940930 Oh et al. Aug 1999 A
5942869 Katou et al. Aug 1999 A
5943730 Boomgaarden Aug 1999 A
5943733 Tagliaferri Aug 1999 A
5943933 Evans et al. Aug 1999 A
5947225 Kawakami et al. Sep 1999 A
5950408 Schaedler Sep 1999 A
5959423 Nakanishi et al. Sep 1999 A
5968281 Wright et al. Oct 1999 A
5974348 Rocks Oct 1999 A
5974365 Mitchell Oct 1999 A
5983448 Wright et al. Nov 1999 A
5984880 Lander et al. Nov 1999 A
5989700 Krivopal Nov 1999 A
5991951 Kubo et al. Nov 1999 A
5995883 Nishikado Nov 1999 A
5995884 Allen et al. Nov 1999 A
5996167 Close Dec 1999 A
5998953 Nakamura et al. Dec 1999 A
5998971 Corbridge Dec 1999 A
6000088 Wright et al. Dec 1999 A
6009358 Angott et al. Dec 1999 A
6012618 Matsuo Jan 2000 A
6021545 Delgado et al. Feb 2000 A
6023813 Thatcher et al. Feb 2000 A
6023814 Imamura Feb 2000 A
6025687 Himeda et al. Feb 2000 A
6026539 Mouw et al. Feb 2000 A
6030464 Azevedo Feb 2000 A
6030465 Marcussen et al. Feb 2000 A
6032327 Oka et al. Mar 2000 A
6032542 Warnick et al. Mar 2000 A
6036572 Sze Mar 2000 A
6038501 Kawakami Mar 2000 A
6040669 Hog Mar 2000 A
6041471 Charky et al. Mar 2000 A
6041472 Kasen et al. Mar 2000 A
6046800 Ohtomo et al. Apr 2000 A
6049620 Dickinson et al. Apr 2000 A
6050648 Keleny Apr 2000 A
6052821 Chouly et al. Apr 2000 A
6055042 Sarangapani Apr 2000 A
6055702 Imamura et al. May 2000 A
6061868 Moritsch et al. May 2000 A
6065182 Wright et al. May 2000 A
6070290 Schwarze et al. Jun 2000 A
6073432 Schaedler Jun 2000 A
6076025 Ueno et al. Jun 2000 A
6076026 Jambhekar et al. Jun 2000 A
6076226 Reed Jun 2000 A
6076227 Schallig et al. Jun 2000 A
6081257 Zeller Jun 2000 A
6088020 Mor Jul 2000 A
6094775 Behmer Aug 2000 A
6099091 Campbell Aug 2000 A
6101671 Wright et al. Aug 2000 A
6108031 King et al. Aug 2000 A
6108067 Okamoto Aug 2000 A
6108076 Hanseder Aug 2000 A
6108269 Kabel Aug 2000 A
6108597 Kirchner et al. Aug 2000 A
6108859 Burgoon Aug 2000 A
6112143 Allen et al. Aug 2000 A
6112996 Matsuo Sep 2000 A
6119057 Kawagoe Sep 2000 A
6122798 Kobayashi et al. Sep 2000 A
6124694 Bancroft et al. Sep 2000 A
6125498 Roberts et al. Oct 2000 A
6131237 Kasper et al. Oct 2000 A
6138063 Himeda Oct 2000 A
6142252 Kinto et al. Nov 2000 A
6146041 Chen et al. Nov 2000 A
6146278 Kobayashi Nov 2000 A
6154279 Thayer Nov 2000 A
6154694 Aoki et al. Nov 2000 A
6160479 Åhlén Dec 2000 A
6167332 Kurtzberg et al. Dec 2000 A
6167587 Kasper et al. Jan 2001 B1
6192548 Huffman Feb 2001 B1
6192549 Kasen et al. Feb 2001 B1
6202243 Beaufoy et al. Mar 2001 B1
6216307 Kaleta et al. Apr 2001 B1
6220865 Macri et al. Apr 2001 B1
6226830 Hendriks et al. May 2001 B1
6230362 Kasper et al. May 2001 B1
6237741 Guidetti May 2001 B1
6240342 Fiegert et al. May 2001 B1
6243913 Frank et al. Jun 2001 B1
6255793 Peless et al. Jul 2001 B1
6259979 Holmquist Jul 2001 B1
6261379 Conrad et al. Jul 2001 B1
6263539 Baig Jul 2001 B1
6263989 Won Jul 2001 B1
6272936 Oreper et al. Aug 2001 B1
6276478 Hopkins et al. Aug 2001 B1
6278918 Dickson et al. Aug 2001 B1
6279196 Kasen et al. Aug 2001 B2
6282526 Ganesh Aug 2001 B1
6283034 Miles Sep 2001 B1
6285778 Nakajima et al. Sep 2001 B1
6285930 Dickson et al. Sep 2001 B1
6286181 Kasper et al. Sep 2001 B1
6300737 Begvall et al. Oct 2001 B1
6321337 Reshef et al. Nov 2001 B1
6321515 Colens Nov 2001 B1
6323570 Nishimura et al. Nov 2001 B1
6324714 Walz et al. Dec 2001 B1
6327741 Reed Dec 2001 B1
6332400 Meyer Dec 2001 B1
6339735 Peless et al. Jan 2002 B1
6362875 Burkley Mar 2002 B1
6370453 Sommer Apr 2002 B2
6374155 Wallach et al. Apr 2002 B1
6374157 Takamura Apr 2002 B1
6381802 Park May 2002 B2
6385515 Dickson et al. May 2002 B1
6388013 Saraf et al. May 2002 B1
6389329 Colens May 2002 B1
6397429 Legatt et al. Jun 2002 B1
6400048 Nishimura et al. Jun 2002 B1
6401294 Kasper Jun 2002 B2
6408226 Byrne et al. Jun 2002 B1
6412141 Kasper et al. Jul 2002 B2
6415203 Inoue et al. Jul 2002 B1
6418586 Fulghum Jul 2002 B2
6421870 Basham et al. Jul 2002 B1
6427285 Legatt et al. Aug 2002 B1
6430471 Kintou et al. Aug 2002 B1
6431296 Won Aug 2002 B1
6437227 Theimer Aug 2002 B1
6437465 Nishimura et al. Aug 2002 B1
6438456 Feddema et al. Aug 2002 B1
6438793 Miner et al. Aug 2002 B1
6442476 Poropat Aug 2002 B1
6442789 Legatt et al. Sep 2002 B1
6443509 Levin et al. Sep 2002 B1
6444003 Sutcliffe Sep 2002 B1
6446302 Kasper et al. Sep 2002 B1
6454036 Airey et al. Sep 2002 B1
D464091 Christianson Oct 2002 S
6457206 Judson Oct 2002 B1
6459955 Bartsch et al. Oct 2002 B1
6463368 Feiten et al. Oct 2002 B1
6465982 Bergvall et al. Oct 2002 B1
6473167 Odell Oct 2002 B1
6480762 Uchikubo et al. Nov 2002 B1
6481515 Kirkpatrick et al. Nov 2002 B1
6490539 Dickson et al. Dec 2002 B1
6491127 Holmberg et al. Dec 2002 B1
6493612 Bisset et al. Dec 2002 B1
6493613 Peless et al. Dec 2002 B2
6496754 Song et al. Dec 2002 B2
6496755 Wallach et al. Dec 2002 B2
6502657 Kerrebrock et al. Jan 2003 B2
6504610 Bauer et al. Jan 2003 B1
6507773 Parker et al. Jan 2003 B2
6519808 Legatt et al. Feb 2003 B2
6525509 Petersson et al. Feb 2003 B1
D471243 Cioffi et al. Mar 2003 S
6530102 Pierce et al. Mar 2003 B1
6530117 Peterson Mar 2003 B2
6532404 Colens Mar 2003 B2
6535793 Allard Mar 2003 B2
6540424 Hall et al. Apr 2003 B1
6540607 Mokris et al. Apr 2003 B2
6543210 Rostoucher et al. Apr 2003 B2
6548982 Papanikolopoulos et al. Apr 2003 B1
6553612 Dyson et al. Apr 2003 B1
6556722 Russell et al. Apr 2003 B1
6556892 Kuroki et al. Apr 2003 B2
6557104 Vu et al. Apr 2003 B2
D474312 Stephens et al. May 2003 S
6563130 Dworkowski et al. May 2003 B2
6571415 Gerber et al. Jun 2003 B2
6571422 Gordon et al. Jun 2003 B1
6572711 Sclafani et al. Jun 2003 B2
6574536 Kawagoe et al. Jun 2003 B1
6580246 Jacobs Jun 2003 B2
6581239 Dyson et al. Jun 2003 B1
6584376 Van Kommer Jun 2003 B1
6586908 Petersson et al. Jul 2003 B2
6587573 Stam et al. Jul 2003 B1
6590222 Bisset et al. Jul 2003 B1
6594551 McKinney et al. Jul 2003 B2
6594844 Jones Jul 2003 B2
D478884 Slipy et al. Aug 2003 S
6601265 Burlington Aug 2003 B1
6604021 Imai et al. Aug 2003 B2
6604022 Parker Aug 2003 B2
6605156 Clark et al. Aug 2003 B1
6609269 Kasper Aug 2003 B2
6611120 Song et al. Aug 2003 B2
6611734 Parker et al. Aug 2003 B2
6611738 Ruffner Aug 2003 B2
6615108 Peless et al. Sep 2003 B1
6615434 Davis et al. Sep 2003 B1
6615885 Ohm Sep 2003 B1
6622465 Jerome et al. Sep 2003 B2
6624744 Wilson et al. Sep 2003 B1
6625843 Kim et al. Sep 2003 B2
6629028 Paromtchik et al. Sep 2003 B2
6633150 Wallach et al. Oct 2003 B1
6637546 Wang Oct 2003 B1
6639659 Granger Oct 2003 B2
6658325 Zweig Dec 2003 B2
6658354 Lin Dec 2003 B2
6658692 Lenkiewicz et al. Dec 2003 B2
6658693 Reed, Jr. Dec 2003 B1
6661239 Ozick Dec 2003 B1
6662889 De Fazio et al. Dec 2003 B2
6668951 Won Dec 2003 B2
6670817 Fournier et al. Dec 2003 B2
6671592 Bisset et al. Dec 2003 B1
6671925 Field et al. Jan 2004 B2
6677938 Maynard Jan 2004 B1
6687571 Byrne et al. Feb 2004 B1
6690134 Jones et al. Feb 2004 B1
6690993 Foulke et al. Feb 2004 B2
6697147 Ko et al. Feb 2004 B2
6705332 Field et al. Mar 2004 B2
6711280 Stafsudd et al. Mar 2004 B2
6732826 Song et al. May 2004 B2
6735811 Field et al. May 2004 B2
6735812 Hekman et al. May 2004 B2
6737591 Lapstun et al. May 2004 B1
6741054 Koselka et al. May 2004 B2
6741364 Lange et al. May 2004 B2
6748297 Song et al. Jun 2004 B2
6756703 Chang Jun 2004 B2
6760647 Nourbakhsh et al. Jul 2004 B2
6764373 Osawa et al. Jul 2004 B1
6769004 Barrett Jul 2004 B2
6774596 Bisset Aug 2004 B1
6779380 Nieuwkamp Aug 2004 B1
6781338 Jones et al. Aug 2004 B2
6809490 Jones et al. Oct 2004 B2
6810305 Kirkpatrick Oct 2004 B2
6810350 Blakley Oct 2004 B2
6830120 Yashima et al. Dec 2004 B1
6832407 Salem et al. Dec 2004 B2
6836701 McKee Dec 2004 B2
6841963 Song et al. Jan 2005 B2
6845297 Allard Jan 2005 B2
6848146 Wright et al. Feb 2005 B2
6854148 Rief et al. Feb 2005 B1
6856811 Burdue et al. Feb 2005 B2
6859010 Jeon et al. Feb 2005 B2
6859682 Naka et al. Feb 2005 B2
6860206 Rudakevych et al. Mar 2005 B1
6865447 Lau et al. Mar 2005 B2
6870792 Chiappetta Mar 2005 B2
6871115 Huang et al. Mar 2005 B2
6883201 Jones et al. Apr 2005 B2
6886651 Slocum et al. May 2005 B1
6888333 Laby May 2005 B2
6901624 Mori et al. Jun 2005 B2
6906702 Tanaka et al. Jun 2005 B1
6101670 Song Jul 2005 A1
6914403 Tsurumi Jul 2005 B2
6917854 Bayer Jul 2005 B2
6925357 Wang et al. Aug 2005 B2
6925679 Wallach et al. Aug 2005 B2
6929548 Wang Aug 2005 B2
D510066 Hickey et al. Sep 2005 S
6938298 Aasen Sep 2005 B2
6940291 Ozick Sep 2005 B1
6941199 Bottomley et al. Sep 2005 B1
6956348 Landry et al. Oct 2005 B2
6957712 Song et al. Oct 2005 B2
6960986 Asama et al. Nov 2005 B2
6965209 Jones et al. Nov 2005 B2
6965211 Tsurumi Nov 2005 B2
6968592 Takeuchi et al. Nov 2005 B2
6971140 Kim Dec 2005 B2
6975246 Trudeau Dec 2005 B1
6980229 Ebersole Dec 2005 B1
6985556 Shanmugavel et al. Jan 2006 B2
6993954 George et al. Feb 2006 B1
6999850 McDonald Feb 2006 B2
7013527 Thomas et al. Mar 2006 B2
7024278 Chiappetta et al. Apr 2006 B2
7024280 Parker et al. Apr 2006 B2
7027893 Perry et al. Apr 2006 B2
7030768 Wanie Apr 2006 B2
7031805 Lee et al. Apr 2006 B2
7032469 Bailey Apr 2006 B2
7040869 Beenker May 2006 B2
7051399 Field et al. May 2006 B2
7053578 Diehl et al. May 2006 B2
7054716 McKee et al. May 2006 B2
5987383 Keller et al. Jun 2006 A1
7055210 Keppler et al. Jun 2006 B2
7057120 Ma et al. Jun 2006 B2
7057643 Iida et al. Jun 2006 B2
7059012 Song et al. Jun 2006 B2
7065430 Naka et al. Jun 2006 B2
7066291 Martins et al. Jun 2006 B2
7069124 Whittaker et al. Jun 2006 B1
7079923 Abramson et al. Jul 2006 B2
7085623 Siegers Aug 2006 B2
7085624 Aldred et al. Aug 2006 B2
7113847 Chmura et al. Sep 2006 B2
7133746 Abramson et al. Nov 2006 B2
7142198 Lee Nov 2006 B2
7148458 Schell et al. Dec 2006 B2
7155308 Jones Dec 2006 B2
7167775 Abramson et al. Jan 2007 B2
7171285 Kim et al. Jan 2007 B2
7173391 Jones et al. Feb 2007 B2
7174238 Zweig Feb 2007 B1
7188000 Chiappetta et al. Mar 2007 B2
7193384 Norman et al. Mar 2007 B1
7196487 Jones et al. Mar 2007 B2
7201786 Wegelin et al. Apr 2007 B2
7206677 Huldén Apr 2007 B2
7211980 Bruemmer et al. May 2007 B1
7225500 Diehl et al. Jun 2007 B2
7246405 Yan Jul 2007 B2
7248951 Huldén Jul 2007 B2
7275280 Haegermarck et al. Oct 2007 B2
7283892 Boillot et al. Oct 2007 B1
7288912 Landry et al. Oct 2007 B2
7318248 Yan Jan 2008 B1
7320149 Huffman et al. Jan 2008 B1
7321807 Laski Jan 2008 B2
7324870 Lee Jan 2008 B2
7328196 Peters Feb 2008 B2
7332890 Cohen et al. Feb 2008 B2
7346428 Huffman et al. Mar 2008 B1
7352153 Yan Apr 2008 B2
7359766 Jeon et al. Apr 2008 B2
7360277 Moshenrose et al. Apr 2008 B2
7363108 Noda et al. Apr 2008 B2
7388879 Sabe et al. Jun 2008 B2
7389156 Ziegler et al. Jun 2008 B2
7389166 Harwig et al. Jun 2008 B2
7408157 Yan Aug 2008 B2
7418762 Arai et al. Sep 2008 B2
7430455 Casey et al. Sep 2008 B2
7430462 Chiu et al. Sep 2008 B2
7441298 Svendsen et al. Oct 2008 B2
7444206 Abramson et al. Oct 2008 B2
7448113 Jones et al. Nov 2008 B2
7459871 Landry et al. Dec 2008 B2
7467026 Sakagami et al. Dec 2008 B2
7474941 Kim et al. Jan 2009 B2
7503096 Lin Mar 2009 B2
7515991 Egawa et al. Apr 2009 B2
7539557 Yamauchi May 2009 B2
7555363 Augenbraun et al. Jun 2009 B2
7557703 Yamada et al. Jul 2009 B2
7568259 Yan Aug 2009 B2
7571511 Jones et al. Aug 2009 B2
7578020 Jaworski et al. Aug 2009 B2
7600521 Woo Oct 2009 B2
7603744 Reindle Oct 2009 B2
7611583 Buckley et al. Nov 2009 B2
7617557 Reindle Nov 2009 B2
7620476 Morse et al. Nov 2009 B2
7636928 Uno Dec 2009 B2
7636982 Jones et al. Dec 2009 B2
7647144 Haegermarck Jan 2010 B2
7650666 Jang Jan 2010 B2
7660650 Kawagoe et al. Feb 2010 B2
7663333 Jones et al. Feb 2010 B2
7693605 Park Apr 2010 B2
7706917 Chiappetta et al. Apr 2010 B1
7761954 Ziegler et al. Jul 2010 B2
7765635 Park Aug 2010 B2
7784147 Burkholder et al. Aug 2010 B2
7801645 Taylor et al. Sep 2010 B2
7805220 Taylor et al. Sep 2010 B2
7809944 Kawamoto Oct 2010 B2
7832048 Harwig et al. Nov 2010 B2
7849555 Hahm et al. Dec 2010 B2
7853645 Brown et al. Dec 2010 B2
7920941 Park et al. Apr 2011 B2
7937800 Yan May 2011 B2
7957836 Myeong et al. Jun 2011 B2
8087117 Kapoor et al. Jan 2012 B2
20010004719 Sommer Jun 2001 A1
20010013929 Torsten Aug 2001 A1
20010020200 Das et al. Sep 2001 A1
20010025183 Shahidi Sep 2001 A1
20010037163 Allard Nov 2001 A1
20010043509 Green et al. Nov 2001 A1
20010045883 Holdaway et al. Nov 2001 A1
20010047231 Peless et al. Nov 2001 A1
20010047895 De et al. Dec 2001 A1
20020011367 Kolesnik Jan 2002 A1
20020011813 Koselka et al. Jan 2002 A1
20020016649 Jones Feb 2002 A1
20020021219 Edwards Feb 2002 A1
20020027652 Paromtchik et al. Mar 2002 A1
20020036779 Kiyoi et al. Mar 2002 A1
20020081937 Yamada et al. Jun 2002 A1
20020095239 Wallach et al. Jul 2002 A1
20020097400 Jung et al. Jul 2002 A1
20020104963 Mancevski Aug 2002 A1
20020108209 Peterson Aug 2002 A1
20020112742 Bredo et al. Aug 2002 A1
20020113973 Ge Aug 2002 A1
20020116089 Kirkpatrick Aug 2002 A1
20020120364 Colens Aug 2002 A1
20020124343 Reed Sep 2002 A1
20020153185 Song et al. Oct 2002 A1
20020156556 Ruffner Oct 2002 A1
20020159051 Guo Oct 2002 A1
20020166193 Kasper Nov 2002 A1
20020169521 Goodman et al. Nov 2002 A1
20020173877 Zweig Nov 2002 A1
20020189871 Won Dec 2002 A1
20030009259 Hattori et al. Jan 2003 A1
20030019071 Field et al. Jan 2003 A1
20030023356 Keable Jan 2003 A1
20030024986 Mazz et al. Feb 2003 A1
20030025472 Jones et al. Feb 2003 A1
20030028286 Glenn et al. Feb 2003 A1
20030030399 Jacobs Feb 2003 A1
20030058262 Sato et al. Mar 2003 A1
20030060928 Abramson et al. Mar 2003 A1
20030067451 Tagg et al. Apr 2003 A1
20030097875 Lentz et al. May 2003 A1
20030120389 Abramson et al. Jun 2003 A1
20030124312 Autumn Jul 2003 A1
20030126352 Barrett Jul 2003 A1
20030137268 Papanikolopoulos et al. Jul 2003 A1
20030146384 Logsdon et al. Aug 2003 A1
20030159232 Hekman et al. Aug 2003 A1
20030168081 Lee et al. Sep 2003 A1
20030175138 Beenker Sep 2003 A1
20030192144 Song et al. Oct 2003 A1
20030193657 Uomori et al. Oct 2003 A1
20030216834 Allard Nov 2003 A1
20030221114 Hino et al. Nov 2003 A1
20030229421 Chmura et al. Dec 2003 A1
20030229474 Suzuki et al. Dec 2003 A1
20030233171 Heiligensetzer Dec 2003 A1
20030233177 Johnson et al. Dec 2003 A1
20030233870 Mancevski Dec 2003 A1
20030233930 Ozick Dec 2003 A1
20040016077 Song et al. Jan 2004 A1
20040020000 Jones Feb 2004 A1
20040030448 Solomon Feb 2004 A1
20040030449 Solomon Feb 2004 A1
20040030450 Solomon Feb 2004 A1
20040030451 Solomon Feb 2004 A1
20040030570 Solomon Feb 2004 A1
20040030571 Solomon Feb 2004 A1
20040031113 Wosewick et al. Feb 2004 A1
20040049877 Jones et al. Mar 2004 A1
20040055163 McCambridge et al. Mar 2004 A1
20040068351 Solomon Apr 2004 A1
20040068415 Solomon Apr 2004 A1
20040068416 Solomon Apr 2004 A1
20040074038 Im et al. Apr 2004 A1
20040074044 Diehl et al. Apr 2004 A1
20040076324 Burl et al. Apr 2004 A1
20040083570 Song et al. May 2004 A1
20040085037 Jones et al. May 2004 A1
20040088079 Lavarec et al. May 2004 A1
20040093122 Galibraith May 2004 A1
20040098167 Yi et al. May 2004 A1
20040111184 Chiappetta et al. Jun 2004 A1
20040111821 Lenkiewicz et al. Jun 2004 A1
20040113777 Matsuhira et al. Jun 2004 A1
20040117064 McDonald Jun 2004 A1
20040117846 Karaoguz et al. Jun 2004 A1
20040118998 Wingett et al. Jun 2004 A1
20040128028 Miyamoto et al. Jul 2004 A1
20040133316 Dean Jul 2004 A1
20040134336 Solomon Jul 2004 A1
20040134337 Solomon Jul 2004 A1
20040143919 Wilder Jul 2004 A1
20040148419 Chen et al. Jul 2004 A1
20040148731 Damman et al. Aug 2004 A1
20040153212 Profio et al. Aug 2004 A1
20040156541 Jeon et al. Aug 2004 A1
20040158357 Lee et al. Aug 2004 A1
20040181706 Chen et al. Sep 2004 A1
20040187249 Jones et al. Sep 2004 A1
20040187457 Colens Sep 2004 A1
20040196451 Aoyama Oct 2004 A1
20040200505 Taylor et al. Oct 2004 A1
20040204792 Taylor et al. Oct 2004 A1
20040204804 Lee et al. Oct 2004 A1
20040210345 Noda et al. Oct 2004 A1
20040210347 Sawada et al. Oct 2004 A1
20040211444 Taylor et al. Oct 2004 A1
20040221790 Sinclair et al. Nov 2004 A1
20040236468 Taylor et al. Nov 2004 A1
20040244138 Taylor et al. Dec 2004 A1
20040255425 Arai et al. Dec 2004 A1
20050000543 Taylor et al. Jan 2005 A1
20050010330 Abramson et al. Jan 2005 A1
20050010331 Taylor et al. Jan 2005 A1
20050015913 Kim et al. Jan 2005 A1
20050021181 Kim et al. Jan 2005 A1
20050028316 Thomas et al. Feb 2005 A1
20050053912 Roth et al. Mar 2005 A1
20050055796 Wright et al. Mar 2005 A1
20050067994 Jones et al. Mar 2005 A1
20050085947 Aldred et al. Apr 2005 A1
20050091782 Gordon et al. May 2005 A1
20050091786 Wright et al. May 2005 A1
20050137749 Jeon et al. Jun 2005 A1
20050144751 Kegg et al. Jul 2005 A1
20050150074 Diehl et al. Jul 2005 A1
20050150519 Keppler et al. Jul 2005 A1
20050154795 Kuz et al. Jul 2005 A1
20050156562 Cohen et al. Jul 2005 A1
20050163119 Ito et al. Jul 2005 A1
20050165508 Kanda et al. Jul 2005 A1
20050166354 Uehigashi Aug 2005 A1
20050166355 Tani Aug 2005 A1
20050172445 Diehl et al. Aug 2005 A1
20050183229 Uehigashi Aug 2005 A1
20050183230 Uehigashi Aug 2005 A1
20050187678 Myeong et al. Aug 2005 A1
20050192707 Park et al. Sep 2005 A1
20050204717 Colens Sep 2005 A1
20050209736 Kawagoe Sep 2005 A1
20050211880 Schell et al. Sep 2005 A1
20050212929 Schell et al. Sep 2005 A1
20050213082 DiBernardo et al. Sep 2005 A1
20050213109 Schell et al. Sep 2005 A1
20050217042 Reindle Oct 2005 A1
20050218852 Landry et al. Oct 2005 A1
20050222933 Wesby Oct 2005 A1
20050229340 Sawalski et al. Oct 2005 A1
20050229355 Crouch et al. Oct 2005 A1
20050235451 Yan Oct 2005 A1
20050251292 Casey et al. Nov 2005 A1
20050255425 Pierson Nov 2005 A1
20050258154 Blankenship et al. Nov 2005 A1
20050273967 Taylor et al. Dec 2005 A1
20050288819 de Guzman Dec 2005 A1
20060000050 Cipolla et al. Jan 2006 A1
20060009879 Lynch et al. Jan 2006 A1
20060010638 Shimizu et al. Jan 2006 A1
20060020369 Taylor et al. Jan 2006 A1
20060020370 Abramson Jan 2006 A1
20060021168 Nishikawa Feb 2006 A1
20060025134 Cho et al. Feb 2006 A1
20060037170 Shimizu Feb 2006 A1
20060042042 Mertes et al. Mar 2006 A1
20060044546 Lewin et al. Mar 2006 A1
20060060216 Woo Mar 2006 A1
20060061657 Rew et al. Mar 2006 A1
20060064828 Stein et al. Mar 2006 A1
20060087273 Ko et al. Apr 2006 A1
20060089765 Pack et al. Apr 2006 A1
20060100741 Jung May 2006 A1
20060107894 Buckley et al. May 2006 A1
20060119839 Bertin et al. Jun 2006 A1
20060143295 Costa et al. Jun 2006 A1
20060146776 Kim Jul 2006 A1
20060150361 Aldred et al. Jul 2006 A1
20060184293 Konandreas et al. Aug 2006 A1
20060185690 Song et al. Aug 2006 A1
20060190133 Konandreas et al. Aug 2006 A1
20060190134 Ziegler et al. Aug 2006 A1
20060190146 Morse et al. Aug 2006 A1
20060196003 Song et al. Sep 2006 A1
20060200281 Ziegler et al. Sep 2006 A1
20060220900 Ceskutti et al. Oct 2006 A1
20060229774 Park et al. Oct 2006 A1
20060259194 Chiu Nov 2006 A1
20060259494 Watson et al. Nov 2006 A1
20060278161 Burkholder et al. Dec 2006 A1
20060288519 Jaworski et al. Dec 2006 A1
20060293787 Kanda et al. Dec 2006 A1
20060293808 Qian Dec 2006 A1
20070006404 Cheng et al. Jan 2007 A1
20070016328 Ziegler et al. Jan 2007 A1
20070017061 Yan Jan 2007 A1
20070028574 Yan Feb 2007 A1
20070032904 Kawagoe et al. Feb 2007 A1
20070042716 Goodall et al. Feb 2007 A1
20070043459 Abbott et al. Feb 2007 A1
20070061041 Zweig Mar 2007 A1
20070061043 Ermakov et al. Mar 2007 A1
20070114975 Cohen et al. May 2007 A1
20070142964 Abramson Jun 2007 A1
20070150096 Yeh et al. Jun 2007 A1
20070156286 Yamauchi Jul 2007 A1
20070157415 Lee et al. Jul 2007 A1
20070157420 Lee et al. Jul 2007 A1
20070179670 Chiappetta et al. Aug 2007 A1
20070226949 Hahm et al. Oct 2007 A1
20070234492 Svendsen et al. Oct 2007 A1
20070244610 Ozick et al. Oct 2007 A1
20070245511 Hahm et al. Oct 2007 A1
20070250212 Halloran et al. Oct 2007 A1
20070261193 Gordon et al. Nov 2007 A1
20070266508 Jones et al. Nov 2007 A1
20080007203 Cohen et al. Jan 2008 A1
20080039974 Sandin et al. Feb 2008 A1
20080052846 Kapoor et al. Mar 2008 A1
20080091304 Ozick et al. Apr 2008 A1
20080109126 Sandin et al. May 2008 A1
20080134458 Ziegler et al. Jun 2008 A1
20080140255 Ziegler et al. Jun 2008 A1
20080155768 Ziegler et al. Jul 2008 A1
20080184518 Taylor et al. Aug 2008 A1
20080276407 Schnittman et al. Nov 2008 A1
20080281470 Gilbert et al. Nov 2008 A1
20080282494 Won et al. Nov 2008 A1
20080294288 Yamauchi Nov 2008 A1
20080302586 Yan Dec 2008 A1
20080307590 Jones et al. Dec 2008 A1
20090007366 Svendsen et al. Jan 2009 A1
20090038089 Landry et al. Feb 2009 A1
20090048727 Hong et al. Feb 2009 A1
20090049640 Lee et al. Feb 2009 A1
20090055022 Casey et al. Feb 2009 A1
20090102296 Greene et al. Apr 2009 A1
20090292393 Casey et al. Nov 2009 A1
20100006028 Buckley et al. Jan 2010 A1
20100011529 Won et al. Jan 2010 A1
20100049365 Jones et al. Feb 2010 A1
20100063628 Landry et al. Mar 2010 A1
20100107355 Won et al. May 2010 A1
20100257690 Jones et al. Oct 2010 A1
20100257691 Jones et al. Oct 2010 A1
20100263158 Jones et al. Oct 2010 A1
20100268384 Jones et al. Oct 2010 A1
20100293742 Chung et al. Nov 2010 A1
20100312429 Jones et al. Dec 2010 A1
RE28268 Dec 1974 E1
Foreign Referenced Citations (479)
Number Date Country
2003275566 Jun 2004 AU
2128842 Dec 1980 DE
3317376 Nov 1984 DE
3536907 Feb 1989 DE
3404202 Dec 1992 DE
199311014 Oct 1993 DE
4414683 Oct 1995 DE
4338841 Aug 1999 DE
19849978 Feb 2001 DE
19849978 Feb 2001 DE
10242257 Apr 2003 DE
102004038074 Jun 2005 DE
10357636 Jul 2005 DE
102004041021 Aug 2005 DE
102005046813 Apr 2007 DE
198803389 Dec 1988 DK
114926 Aug 1984 EP
265542 May 1988 EP
281085 Sep 1988 EP
286328 Oct 1988 EP
352045 Jan 1990 EP
307381 Jul 1990 EP
358628 May 1991 EP
437024 Jul 1991 EP
389459 Oct 1991 EP
433697 Dec 1992 EP
479273 May 1993 EP
294101 Dec 1993 EP
554978 Mar 1994 EP
615719 Sep 1994 EP
0 792 726 Sep 1997 EP
861629 Sep 1998 EP
930040 Oct 1999 EP
845237 Apr 2000 EP
1018315 Jul 2000 EP
1172719 Jan 2002 EP
1228734 Jun 2003 EP
1 331 537 Jul 2003 EP
1 331 537 Jul 2003 EP
1331537 Jul 2003 EP
1139847 Aug 2003 EP
1380245 Jan 2004 EP
1380246 Mar 2005 EP
1553472 Jul 2005 EP
1557730 Jul 2005 EP
1642522 Nov 2007 EP
1672455 Dec 2007 EP
1836941 Oct 2009 EP
2238196 Nov 2006 ES
722755 Mar 1932 FR
2601443 Nov 1991 FR
2 828 589 Aug 2001 FR
2828589 Feb 2003 FR
702426 Jan 1954 GB
2128842 Apr 1986 GB
2213047 Aug 1989 GB
2225221 May 1990 GB
2 283 838 May 1995 GB
2284957 Jun 1995 GB
2267360 Dec 1995 GB
2300082 Sep 1999 GB
2344747 May 2002 GB
2404330 Jul 2005 GB
2409966 Jul 2005 GB
2417354 Feb 2006 GB
53021869 Feb 1978 JP
53110257 Sep 1978 JP
53110257 Sep 1978 JP
57014726 Jan 1982 JP
57064217 Apr 1982 JP
59005315 Jan 1984 JP
59005315 Feb 1984 JP
59033511 Mar 1984 JP
59094005 May 1984 JP
59099308 Jul 1984 JP
59112311 Jul 1984 JP
59033511 Aug 1984 JP
59120124 Aug 1984 JP
59131668 Sep 1984 JP
59164973 Sep 1984 JP
59184917 Oct 1984 JP
2283343 Nov 1984 JP
59212924 Dec 1984 JP
59226909 Dec 1984 JP
60089213 May 1985 JP
60089213 Jun 1985 JP
60211510 Oct 1985 JP
60259895 Dec 1985 JP
61023221 Jan 1986 JP
61097712 May 1986 JP
61023221 Jun 1986 JP
61160366 Jul 1986 JP
62074018 Apr 1987 JP
62070709 May 1987 JP
62-120510 Jun 1987 JP
62-154008 Jul 1987 JP
62154008 Jul 1987 JP
62164431 Oct 1987 JP
62263507 Nov 1987 JP
62263508 Nov 1987 JP
62189057 Dec 1987 JP
63079623 Apr 1988 JP
63-183032 Jul 1988 JP
63158032 Jul 1988 JP
63203483 Aug 1988 JP
63-241610 Oct 1988 JP
63241610 Oct 1988 JP
1118752 May 1989 JP
2-6312 Jan 1990 JP
2006312 Jan 1990 JP
2283343 Nov 1990 JP
03-051023 Mar 1991 JP
3051023 Mar 1991 JP
3197758 Aug 1991 JP
3201903 Sep 1991 JP
4019586 Mar 1992 JP
4084921 Mar 1992 JP
5023269 Apr 1993 JP
5091604 Apr 1993 JP
5095879 Apr 1993 JP
5042076 Jun 1993 JP
5046246 Jun 1993 JP
5150827 Jun 1993 JP
5150829 Jun 1993 JP
5046239 Jul 1993 JP
5054620 Jul 1993 JP
5060049 Aug 1993 JP
5040519 Oct 1993 JP
5257527 Oct 1993 JP
5257533 Oct 1993 JP
5257533 Oct 1993 JP
5285861 Nov 1993 JP
5302836 Nov 1993 JP
5312514 Nov 1993 JP
5091604 Dec 1993 JP
5341904 Dec 1993 JP
6003251 Jan 1994 JP
6038912 Feb 1994 JP
6137828 May 1994 JP
6154143 Jun 1994 JP
6293095 Oct 1994 JP
06-327598 Nov 1994 JP
6105781 Dec 1994 JP
7047046 Feb 1995 JP
7059702 Mar 1995 JP
07-129239 May 1995 JP
7059702 Jun 1995 JP
7222705 Aug 1995 JP
7222705 Aug 1995 JP
7270518 Oct 1995 JP
7281742 Oct 1995 JP
7281752 Oct 1995 JP
7-295636 Nov 1995 JP
7295636 Nov 1995 JP
7311041 Nov 1995 JP
7311041 Nov 1995 JP
7313417 Dec 1995 JP
7319542 Dec 1995 JP
7319542 Dec 1995 JP
8-16776 Jan 1996 JP
8000393 Jan 1996 JP
8000393 Jan 1996 JP
8016241 Jan 1996 JP
8016241 Feb 1996 JP
8016776 Feb 1996 JP
8063229 Mar 1996 JP
8063229 Mar 1996 JP
8083125 Mar 1996 JP
8083125 Mar 1996 JP
08-089451 Apr 1996 JP
8084696 Apr 1996 JP
8089449 Apr 1996 JP
2520732 May 1996 JP
8123548 May 1996 JP
08-152916 Jun 1996 JP
8152916 Jun 1996 JP
8256960 Oct 1996 JP
8256960 Oct 1996 JP
8263137 Oct 1996 JP
8286741 Nov 1996 JP
8286741 Nov 1996 JP
8286744 Nov 1996 JP
8286744 Nov 1996 JP
8286745 Nov 1996 JP
8286747 Nov 1996 JP
8322774 Dec 1996 JP
8322774 Dec 1996 JP
8335112 Dec 1996 JP
8339297 Dec 1996 JP
9044240 Feb 1997 JP
9047413 Feb 1997 JP
9066855 Mar 1997 JP
9145309 Jun 1997 JP
9160644 Jun 1997 JP
9160644 Jun 1997 JP
9-179625 Jul 1997 JP
9179625 Jul 1997 JP
9179685 Jul 1997 JP
9185410 Jul 1997 JP
9185410 Jul 1997 JP
9192069 Jul 1997 JP
9192069 Jul 1997 JP
9204223 Aug 1997 JP
9204223 Aug 1997 JP
9204224 Aug 1997 JP
9206258 Aug 1997 JP
9206258 Aug 1997 JP
9233712 Sep 1997 JP
09251318 Sep 1997 JP
9251318 Sep 1997 JP
9265319 Oct 1997 JP
9269807 Oct 1997 JP
9269810 Oct 1997 JP
9269824 Oct 1997 JP
02555263 Nov 1997 JP
9319431 Dec 1997 JP
9319431 Dec 1997 JP
9319432 Dec 1997 JP
9319434 Dec 1997 JP
9325812 Dec 1997 JP
10027020 Jan 1998 JP
10055215 Feb 1998 JP
10105233 Apr 1998 JP
10117973 May 1998 JP
10117973 May 1998 JP
10118963 May 1998 JP
10165738 Jun 1998 JP
10177414 Jun 1998 JP
10214114 Aug 1998 JP
10214114 Aug 1998 JP
10228316 Aug 1998 JP
10240342 Sep 1998 JP
10240342 Sep 1998 JP
10240343 Sep 1998 JP
10260727 Sep 1998 JP
10260727 Sep 1998 JP
10295595 Nov 1998 JP
10314088 Dec 1998 JP
11055941 Jan 1999 JP
11065655 Mar 1999 JP
11065655 Mar 1999 JP
11065657 Mar 1999 JP
11085269 Mar 1999 JP
11102219 Apr 1999 JP
11102219 Apr 1999 JP
11102220 Apr 1999 JP
11162454 Jun 1999 JP
11174145 Jul 1999 JP
11175149 Jul 1999 JP
11178764 Jul 1999 JP
11178765 Jul 1999 JP
11-508810 Aug 1999 JP
11212642 Aug 1999 JP
11212642 Aug 1999 JP
11213157 Aug 1999 JP
11508810 Aug 1999 JP
11-510935 Sep 1999 JP
11248806 Sep 1999 JP
11510935 Sep 1999 JP
11282532 Oct 1999 JP
11282533 Oct 1999 JP
11295412 Oct 1999 JP
11346964 Dec 1999 JP
11346964 Dec 1999 JP
2000047728 Feb 2000 JP
2000056006 Feb 2000 JP
2000056831 Feb 2000 JP
2000060782 Feb 2000 JP
2000066722 Mar 2000 JP
2000075925 Mar 2000 JP
2000102499 Apr 2000 JP
10240343 May 2000 JP
03051023 Jun 2000 JP
2000275321 Oct 2000 JP
2000279353 Oct 2000 JP
2000342497 Dec 2000 JP
2000353014 Dec 2000 JP
2000353014 Dec 2000 JP
2001022443 Feb 2001 JP
2001067588 Mar 2001 JP
2001087182 Apr 2001 JP
2001087182 Apr 2001 JP
2001121455 May 2001 JP
2001125641 May 2001 JP
2001508572 Jun 2001 JP
2001197008 Jul 2001 JP
2001197008 Jul 2001 JP
2001216482 Aug 2001 JP
2001-258807 Sep 2001 JP
2001265437 Sep 2001 JP
2001-275908 Oct 2001 JP
2001289939 Oct 2001 JP
2001306170 Nov 2001 JP
2001320781 Nov 2001 JP
2001-525567 Dec 2001 JP
2002-78650 Mar 2002 JP
2002073170 Mar 2002 JP
2002-204768 Jul 2002 JP
22204768 Jul 2002 JP
2002204769 Jul 2002 JP
2002247510 Aug 2002 JP
2002-532178 Oct 2002 JP
3356170 Oct 2002 JP
2002532180 Oct 2002 JP
2002-323925 Nov 2002 JP
3375843 Nov 2002 JP
22323925 Nov 2002 JP
2002333920 Nov 2002 JP
2002-355206 Dec 2002 JP
2002-360471 Dec 2002 JP
2002-360482 Dec 2002 JP
2002360479 Dec 2002 JP
2002366227 Dec 2002 JP
2002369778 Dec 2002 JP
2002369778 Dec 2002 JP
2003-10076 Jan 2003 JP
23028528 Jan 2003 JP
2003010088 Jan 2003 JP
2003015740 Jan 2003 JP
2003028528 Jan 2003 JP
2003-5296 Feb 2003 JP
2003-036116 Feb 2003 JP
2003-38401 Feb 2003 JP
2003-38402 Feb 2003 JP
2003-505127 Feb 2003 JP
23036116 Feb 2003 JP
23052596 Feb 2003 JP
2003036116 Feb 2003 JP
2003047579 Feb 2003 JP
2003-061882 Mar 2003 JP
2003084994 Mar 2003 JP
2003167628 Jun 2003 JP
2003180586 Jul 2003 JP
2003180587 Jul 2003 JP
2003186539 Jul 2003 JP
2003190064 Jul 2003 JP
2003190064 Jul 2003 JP
2003228421 Aug 2003 JP
2003241836 Aug 2003 JP
2003262520 Sep 2003 JP
2003285288 Oct 2003 JP
2003304992 Oct 2003 JP
2003-310489 Nov 2003 JP
2003310509 Nov 2003 JP
2003330543 Nov 2003 JP
2004123040 Apr 2004 JP
2004148021 May 2004 JP
2004160102 Jun 2004 JP
2004166968 Jun 2004 JP
2004174228 Jun 2004 JP
2004198330 Jul 2004 JP
2004219185 Aug 2004 JP
2004351234 Dec 2004 JP
2005118354 May 2005 JP
2005135400 May 2005 JP
2005211360 Aug 2005 JP
2005224265 Aug 2005 JP
2005230032 Sep 2005 JP
2005245916 Sep 2005 JP
2005296511 Oct 2005 JP
2005346700 Dec 2005 JP
2005346700 Dec 2005 JP
2005352707 Dec 2005 JP
2006043071 Feb 2006 JP
2006079145 Mar 2006 JP
2006079157 Mar 2006 JP
2006155274 Jun 2006 JP
2006164223 Jun 2006 JP
2006227673 Aug 2006 JP
2006247467 Sep 2006 JP
2006260161 Sep 2006 JP
2006293662 Oct 2006 JP
2006296697 Nov 2006 JP
2007034866 Feb 2007 JP
2007213180 Aug 2007 JP
04074285 Apr 2008 JP
2009015611 Jan 2009 JP
04300516 Jul 2009 JP
2010198552 Sep 2010 JP
WO 9526512 Oct 1995 WO
WO9526512 Oct 1995 WO
WO9530887 Nov 1995 WO
WO9617258 Feb 1997 WO
WO 9715224 May 1997 WO
WO 9740734 Nov 1997 WO
WO 9741451 Nov 1997 WO
WO9740734 Nov 1997 WO
WO9741451 Nov 1997 WO
WO 9853456 Nov 1998 WO
WO9905580 Feb 1999 WO
WO 9916078 Apr 1999 WO
WO 9928800 Jun 1999 WO
WO 9938056 Jul 1999 WO
WO 9938237 Jul 1999 WO
WO9938056 Jul 1999 WO
WO9938237 Jul 1999 WO
WO 9943250 Sep 1999 WO
WO 9959042 Nov 1999 WO
WO9959042 Nov 1999 WO
WO 0004430 Jan 2000 WO
WO0004430 Jan 2000 WO
WO 0036962 Jun 2000 WO
WO 0038026 Jun 2000 WO
WO 0038029 Jun 2000 WO
WO0038026 Jun 2000 WO
WO0038028 Jun 2000 WO
WO0038029 Jun 2000 WO
WO 0078410 Dec 2000 WO
WO 0106904 Feb 2001 WO
WO 0106905 Feb 2001 WO
WO0106904 Feb 2001 WO
WO0180703 Nov 2001 WO
WO2001080703 Nov 2001 WO
WO0191623 Dec 2001 WO
WO0224292 Mar 2002 WO
WO 0239864 May 2002 WO
WO 0239868 May 2002 WO
WO0239868 May 2002 WO
WO 02058527 Aug 2002 WO
WO 02062194 Aug 2002 WO
W002075350 Sep 2002 WO
WO 02067744 Sep 2002 WO
WO 02067745 Sep 2002 WO
WO 02074150 Sep 2002 WO
WO 02075356 Sep 2002 WO
WO 02075469 Sep 2002 WO
WO 02075470 Sep 2002 WO
WO02067752 Sep 2002 WO
WO02069774 Sep 2002 WO
WO02071175 Sep 2002 WO
WO02075350 Sep 2002 WO
WO02075356 Sep 2002 WO
WO02075469 Sep 2002 WO
WO02075470 Sep 2002 WO
WO 02081074 Oct 2002 WO
WO02081074 Oct 2002 WO
WO 02101477 Dec 2002 WO
WO02101477 Dec 2002 WO
WO03015220 Feb 2003 WO
WO 03026474 Apr 2003 WO
WO 03040845 May 2003 WO
WO 03040846 May 2003 WO
WO02069775 May 2003 WO
WO03040546 May 2003 WO
WO03062850 Jul 2003 WO
WO03062852 Jul 2003 WO
WO 2004006034 Jan 2004 WO
WO 2004004533 Jan 2004 WO
WO2004004534 Jan 2004 WO
WO2004005956 Jan 2004 WO
WO2004006034 Jan 2004 WO
WO 2004058028 Jan 2004 WO
WO 2005077244 Jan 2004 WO
WO 2006068403 Jan 2004 WO
WO2004025947 May 2004 WO
WO2004043215 May 2004 WO
WO2004058028 Jul 2004 WO
WO2004059409 Jul 2004 WO
WO2005006935 Jan 2005 WO
WO2005036292 Apr 2005 WO
WO2005037496 Apr 2005 WO
WO 2005055795 Jun 2005 WO
WO2005055795 Jun 2005 WO
WO2005055796 Jun 2005 WO
WO2005055796 Jun 2005 WO
WO2005076545 Aug 2005 WO
WO2005081074 Sep 2005 WO
WO2005082223 Sep 2005 WO
WO2005083541 Sep 2005 WO
WO2005098475 Oct 2005 WO
WO2005098476 Oct 2005 WO
WO2006046400 May 2006 WO
WO2006061133 Jun 2006 WO
WO2006073248 Jul 2006 WO
WO2006089307 Aug 2006 WO
WO2007028049 Mar 2007 WO
WO2007036490 May 2007 WO
WO2007065033 Jun 2007 WO
WO2007137234 Nov 2007 WO
Non-Patent Literature Citations (234)
Entry
Prassler et al., “A Short History of Cleaning Robots,” Autonomous Robots, vol. 9, pp. 211-226 (2000).
Chamberlin, et al., Robot Locator Beacon System, NASA Goddard SFC, Design Proposal, Feb. 17, 2006.
Cameron Morland, Autonomous Lawnmower Control, Jul. 24, 2002.
Doty, Keith L. et al., “Sweep Strategies for a Sensory-Driven, Behavior-Based Vacuum Cleaning Agent” AAAI 1993 Fall Symposium Series Instantiating Real-World Agents Research Triangle Park, Raleigh, NC, Oct. 22-24, 1993, pp. 1-6.
Electrolux designed for the well-lived home, website: http://www.electroluxusa.com/node57.as[?currentURL =node142,asp%3F, accessed Mar. 18, 2005, 5 pp.
eVac Robotic Vacuum S1727 Instruction Manual, Sharper Image Corp., Copyright 2004, 16 pp.
Everyday Robots, website: http://www.everydayrobots.com/index.php?option=content&task=view&id=9, accessed Apr. 20, 2005, 7 pp.
Facts on the Trilobite webpage: “http://trilobiteelectroluxse/presskit—en/node11335asp?print=yes&pressID=” accessed Dec. 12, 2003, 2 pp.
Friendly Robotics Robotic Vacuum RV400-The Robot Store website: http://www.therobotstore.com/s.nl/ sc.9/category,-109iit.A/id.43/.f, accessed Apr. 20, 2005, 5 pp.
Gat, Erann, Robust Low-computation Sensor-driven Control for Task-Directed Navigation, Proceedings of the 1991 IEEE, International Conference on Robotics and Automation, Sacramento, California, Apr. 1991, pp. 2484-2489.
Hitachi: News release: The home cleaning robot of the autonomous movement type (experimental machine) is developed, website: http://www.i4u.com/japanreleases/hitachirobot.htm., accessed Mar. 18, 2005, 5 pp.
Kärcher Product Manual Download webpage: “http://wwwkarchercom/bta/downloadenshtml?ACTION=SELECTTEILENR&ID=rc3000&submitButtonName=Select+Product+Manual” and associated pdf file “5959-915enpdf (47 MB) English/English” accessed Jan. 21, 2004 (16 pp.).
Karcher RC 3000 Cleaning Robot—user manual Manufacturer: Alfred-Karcher GmbH & Co, Cleaning Systems, Alfred Karcher-Str 28-40, PO Box 160, D-71349 Winnenden, Germany, Dec. 2002.
Kärcher RoboCleaner RC 3000 Product Details webpages: “http://wwwrobocleanerde/english/screen3html” through “. . . screen6html” accessed Dec. 12, 2003 4 pp.
Karcher USA, RC3000 Robotic Cleaner, website: http://www.karcher-usa.com/showproducts.php?op=view—prod&param1=143&param2=&param3=, accessed Mar. 18, 2005, 6 pp.
koolvac Robotic Vacuum Cleaner Owner's Manual, Koolatron, Undated, 26 pp.
NorthStar Low-Cost, Indoor Localization, Evolution robotics, Powering Intelligent Products, 2 pp.
Put Your Roomba . . . On “Automatic” Roomba Timer> Timed Cleaning-Floorvac Robotic Vacuum webpages: http://cgi.ebay.com/ws/eBayISAPI.dll?ViewItem&category=43575198387&rd=1, accessed Apr. 20, 2005, 5 pp.
Put Your Roomba . . . On “Automatic” webpages: “http://www.acomputeredge.com/roomba,” accessed Apr. 20, 2005, 5 pp.
RobotMaid Sweeps Your Floors So You Won't Have to, the Official Site, website: http://www.therobomaid.com/, accessed Mar. 18, 2005, 2 pp.
Robot Review Samsung Robot Vacuum (VC-RP3OW), website: http://www.onrobo.com/reviews/At—Home/ Vacuum—Cleaners/on00vcrp30rosam/index.htm, accessed Mar. 18, 2005, 11 pp.
Robotic Vacuum Cleaner-Blue, website: http://www.sharperimage.com/us/en/catalog/productview.jhtml?sku=S1727BLU, accessed Mar. 18, 2005, 3 pp.
Schofield, Monica. “Neither Master nor Slave” A Practical Study in the Development and Employment of Cleaning Robots, Emerging Technologies and Factory Automation, 1999 Proceedings EFA '99 1999 7th IEEE International Conference on Barcelona, Spain, Oct. 18-21, 1999, pp. 1427-1434.
Sebastian Thrun, Learning Occupancy Grid Maps With Forward Sensor Models, School of Computer Science, Carnegie Mellon University, pp. 1-28.
Wired News: Robot Vacs Are in the House, website: http://www.wired.com/news/print/0,1294,59237,00.html, accessed Mar. 18, 2005, 6 pp.
Zoombot Remote Controlled Vacuum-RV-500 New Roomba 2, website: http://cgi.ebay.com/ws/eBaylSAPl.dll?ViewItem&category=43526&item=4373497618&rd=1, accessed Apr. 20, 2005, 7 pp.
International Preliminary Report for counterpart application, PCT/US2006/046395 dated Feb. 28, 2008.
Borges et al. “Optimal Mobile Robot Pose Estimation Using Geometrical Maps”, IEEE Transactions on Robotics and Automation, vol. 18, No. 1, pp. 87-94, Feb. 2002.
Braunstingl et al. “Fuzzy Logic Wall Following of a Mobile Robot Based on the Concept of General Perception” ICAR '95, 7th International Conference on Advanced Robotics, Sant Feliu De Guixols, Spain, pp. 367-376, Sep. 1995.
Bulusu, et al. “Self Configuring Localization systems: Design and Experimental Evaluation”, ACM Transactions on Embedded Computing Systems vol. 3, No. 1, pp. 24-60, 2003.
Caccia, et al. “Bottom-Following for Remotely Operated Vehicles”, 5th IFAC conference, Alaborg, Denmark, pp. 245-250 Aug. 1, 2000.
Chae, et al. “StarLITE: A new artificial landmark for the navigation of mobile robots”, http://www.irc.atr.jp/jk-nrs2005/pdf/Starlite.pdf, 4 pages, 2005.
Champy “Physical management of IT assets in Data Centers using RFID technologies”, RFID 2005 University, Oct. 12-14, 2005 (NPL0126).
Chiri “Joystick Control for Tiny OS Robot”, http://www.eecs.berkeley.edu/Programs/ugrad/superb/papers2002/chiri.pdf. 12 pages, Aug. 8, 2002.
Christensen et al. “Theoretical Methods for Planning and Control in Mobile Robotics” 1997 First International Conference on Knowledge-Based Intelligent Electronic Systems, Adelaide, Australia, pp. 81-86, May 21-27, 1997.
Andersen et al., “Landmark based navigation strategies”, SPIE Conference on Mobile Robots XIII, SPIE vol. 3525, pp.
Clerentin, et al. “A localization method based on two omnidirectional perception systems cooperation” Proc of IEEE International Conference on Robotics & Automation, San Francisco, CA vol. 2, pp. 1219-1224, Apr. 2000.
Corke “High Performance Visual serving for robots end-point control”. SPIE vol. 2056 Intelligent robots and computer vision 1993.
Cozman et al. “Robot Localization using a Computer Vision Sextant”, IEEE International Midwest Conference on Robotics and Automation, pp. 106-111, 1995.
D'Orazio, et al. “Model based Vision System for mobile robot position estimation”, SPIE vol. 2058 Mobile Robots VIII, pp. 38-49, 1992.
De Bakker, et al. “Smart PSD—array for sheet of light range imaging”, Proc. Of SPIE vol. 3965, pp. 1-12, May 15, 2000.
Desaulniers, et al. “An Efficient Algorithm to find a shortest path for a car-like Robot”, IEEE Transactions on robotics and Automation vol. 11 No. 6, pp. 819-828, Dec. 1995.
Dorfmüller-Ulhaas “Optical Tracking From User Motion to 3D Interaction”, http://www.cg.tuwien.ac.at/research/publications/2002/Dorfmueller-Ulhaas-thesis, 182 pages, 2002.
Dorsch, et al. “Laser Triangulation: Fundamental uncertainty in distance measurement”, Applied Optics, vol. 33 No. 7, pp. 1306-1314, Mar. 1, 1994.
Doty et al. “Sweep Strategies for a Sensory-Driven, Behavior-Based Vacuum Cleaning Agent”, AAAI 1993 Fall Symposium Series, Instantiating Real-World Agents, pp. 1-6, Oct. 22-24, 1993 (NPL0129).
Dudek, et al. “Localizing a Robot with Minimum Travel” Proceedings of the sixth annual ACM-SIAM symposium on Discrete algorithms, vol. 27, No. 2, pp. 583-604, Apr. 1998.
Dulimarta, et al. “Mobile Robot Localization in Indoor Environment”, Pattern Recognition, vol. 30, No. 1, pp. 99-111, 1997.
EBay “Roomba Timer → Timed Cleaning—Floorvac Robotic Vacuum”, Cgi.ebay.com/ws/eBay|SAP|.dll? viewitem&category=43526&item=4375198387&rd=1, 5 pages, Apr. 20, 2005.
Electrolux “Welcome to the Electrolux trilobite” www.electroluxusa.com/node57.asp?currentURL=node142.asp%3F, 2 pages, Mar. 18, 2005.
Eren, et al. “Accuracy in position estimation of mobile robots based on coded infrared signal transmission”, Proceedings: Integrating Intelligent Instrumentation and Control, Instrumentation and Measurement Technology Conference, 1995. IMTC/95. pp. 548-551, 1995.
Eren, et al. “Operation of Mobile Robots in a Structured Infrared Environment”, Proceedings. ‘Sensing, Processing, Networking’, IEEE Instrumentation and Measurement Technology Conference, 1997 (IMTC/97), Ottawa, Canada vol. 1, pp. 20-25, May 19-21, 1997.
Barker, “Navigation by the Stars—Ben Barker 4th Year Project” Power point pp. 1-20.
Becker, et al. “Reliable Navigation Using Landmarks ” IEEE International Conference on Robotics and Automation, 0-7803-1965-6, pp. 401-406, 1995.
Benayad-Cherif, et al., “Mobile Robot Navigation Sensors” SPIE vol. 1831 Mobile Robots, VII, pp. 378-387, 1992.
Facchinetti, Claudio et al. “Using and Learning Vision-Based Self-Positioning for Autonomous Robot Navigation”, ICARCV '94, vol. 3 pp. 1694-1698, 1994.
Betke, et al., “Mobile Robot localization using Landmarks” Proceedings of the IEEE/RSJ/GI International Conference on Intelligent Robots and Systems '94 “Advanced Robotic Systems and the Real World” (IROS '94), vol.
Facchinetti, Claudio et al. “Self-Positioning Robot Navigation Using Ceiling Images Sequences”, ACCV '95, 5 pages, Dec. 5-8, 1995.
Fairfield, Nathaniel et al. “Mobile Robot Localization with Sparse Landmarks”, SPIE vol. 4573 pp. 148-155, 2002.
Favre-Bulle, Bernard “Efficient tracking of 3D—Robot Position by Dynamic Triangulation”, IEEE Instrumentation and Measurement Technology Conference IMTC 98 Session on Instrumentation and Measurement in Robotics, vol. 1, pp. 446-449, May 18-21, 1998.
Fayman “Exploiting Process Integration and Composition in the context of Active Vision”, IEEE Transactions on Systems, Man, and Cybernetics- Part C: Application and reviews, vol. 29 No. 1, pp. 73-86, Feb. 1999.
Florbot GE Plastics Image (1989-1990).
Franz, et al. “Biomimetric robot navigation”, Robotics and Autonomous Systems vol. 30 pp. 133-153, 2000.
Friendly Robotics “Friendly Robotics—Friendly Vac, Robotic Vacuum Cleaner”, www.friendlyrobotics.com/vac.htm. 5 pages, Apr. 20, 2005.
Fuentes, et al. “Mobile Robotics 1994”, University of Rochester. Computer Science Department, TR 588, 44 pages, Dec. 7, 1994.
Bison, P et al., “Using a structured beacon for cooperative position estimation” Robotics and Autonomous Systems vol., 29, No. 1, pp. 33-40, Oct. 1999.
Fukuda, et al. “Navigation System based on Ceiling Landmark Recognition for Autonomous mobile robot”, 1995 IEEE/RSJ International Conference on Intelligent Robots and Systems 95. ‘Human Robot Interaction and Cooperative Robots’, Pittsburgh, PA, pp. 1466/1471, Aug. 5-9, 1995.
Gionis “A hand-held optical surface scanner for environmental Modeling and Virtual Reality”, Virtual Reality World, 16 pages, 1996.
Goncalves et al. “A Visual Front-End for Simultaneous Localization and Mapping”, Proceedings of the 2005 IEEE International Conference on Robotics and Automation, Barcelona, Spain, pp. 44-49, Apr. 2005.
Gregg et al. “Autonomous Lawn Care Applications”, 2006 Florida Conference on Recent Advances in Robotics, FCRAR 2006, pp. 1-5, May 25-26, 2006.
Hamamatsu “SI PIN Diode S5980, S5981 S5870—Multi-element photodiodes for surface mounting”, Hamatsu Photonics, 2 pages, Apr. 2004.
Hammacher Schlemmer “Electrolux Trilobite Robotic Vacuum” www.hammacher.com/publish/71579.asp?promo=xsells, 3 pages, Mar. 18, 2005.
Haralick et al. “Pose Estimation from Corresponding Point Data”, IEEE Transactions on systems, Man, and Cybernetics, vol. 19, No. 6, pp. 1426-1446, Nov. 1989.
Hausler “About the Scaling Behaviour of Optical Range Sensors”, Fringe '97, Proceedings of the 3rd International Workshop on Automatic Processing of Fringe Patterns, Bremen, Germany, pp. 147-155, Sep. 15-17, 1997.
Blaasvaer, et al. “AMOR—An Autonomous Mobile Robot Navigation System”, Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, pp. 2266-2271, 1994.
Hoag, et al. “Navigation and Guidance in interstellar space”, ACTA Astronautica vol. 2, pp. 513-533, Feb. 14, 1975.
Jarosiewicz et al. “Final Report—Lucid”, University of Florida, Departmetn of Electrical and Computer Engineering, EEL 5666—Intelligent Machine Design Laboratory, 50 pages, Aug. 4, 1999.
Jensfelt, et al. “Active Global Localization for a mobile robot using multiple hypothesis tracking”, IEEE Transactions on Robots and Automation vol. 17, No. 5, pp. 748-760, Oct. 2001.
Jeong, et al. “An intelligent map-building system for indoor mobile robot using low cost photo sensors”, SPIE vol. 6042 6 pages, 2005.
Kahney, “Robot Vacs are in the House,” www.wired.com/news/technology/o,1282,59237,00.html, 6 pages, Jun. 18, 2003.
Karcher “Product Manual Download Karch”, www.karcher.com, 17 pages, 2004.
Karcher USA, RC3000 Robotic Cleaner, website: http://www.karcher-usa.com/showproducts.php?op=view prod&paraml=143&param2=&param3=, accessed Mar. 18, 2005, 6 pages.
Karlsson et al., The vSLAM Algorithm for Robust Localization and Mapping, Proceedings of the 2005 IEEE International Conference on Robotics and Automation, Barcelona, Spain, pp. 24-29, Apr. 2005.
Karlsson, et al Core Technologies for service Robotics, IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2004), vol. 3, pp. 2979-2984, Sep. 28-Oct. 2, 2004.
Kleinberg, The Localization Problem for Mobile Robots, Laboratory for Computer Science, Massachusetts Institute of Technology, 1994 IEEE, pp. 521-531, 1994.
Knight, et al., “Localization and Identification of Visual Landmarks”, Journal of Computing Sciences in Colleges, vol. 16 Issue 4, 2001 pp. 312-313, May 2001.
Kolodko et al. “Experimental System for Real-Time Motion Estimation”, Proceedings of the 2003 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM 2003), pp. 981-986, 2003.
Komoriya et al., Planning of Landmark Measurement for the Navigation of a Mobile Robot, Proceedings of the 1992 IEEE/RSJ International Cofnerence on Intelligent Robots and Systems, Raleigh, NC pp. 1476-1481, Jul. 7-10, 1992.
Koolatron “KOOLVAC—Owner's Manual”, 13 pages.
Krotov, et al. “Digital Sextant”, Downloaded from the internet at: http://www.cs.cmu.edu/˜epk/ , 1 page, 1995.
Krupa et al. “Autonomous 3-D Positioning of Surgical Instruments in Robotized Laparoscopic Surgery Using Visual Servoing”, IEEE Transactions on Robotics and Automation, vol. 19, No. 5, pp. 842-853, Oct. 5, 2003.
Kuhl, et al. “Self Localization in Environments using Visual Angles”, VRCAI '04 Proceedings of the 2004 ACM SIGGRAPH international conference on Virtual Reality continuum and its applications in industry, pp. 472-475, 2004.
Kurth, “Range-Only Robot Localization and SLAM with Radio”, http://www.ri.cmu.edu/pub—files/pub4/kurth—derek—2004—1/kurth—derek—2004—1.pdf. 60 pages, May, 2004.
Lambrinos, et al. “A mobile robot employing insect strategies for navigation”, http://www8.cs.umu.se/kurser/TDBD17/VT04/dl/Assignment%20Papers/lambrinos-RAS-2000.pdf, 38 pages, Feb. 19, 1999.
Lang et al. “Visual Measurement of Orientation Using Ceiling Features”, 1994 IEEE, pp. 552-555, 1994.
Lapin, “Adaptive position estimation for an automated guided vehicle”, SPIE vol. 1831 Mobile Robots VII, pp. 82-94, 1992.
LaValle et al. “Robot Motion Planning in a Changing, Partially Predictable Environment”, 1994 IEEE International Symposium on Intelligent Control, Columbus, OH, pp. 261-266, Aug. 16-18, 1994.
Lee, et al. “Localization Of a Mobile Robot Using the Image of a Moving Object”, IEEE Transaction on Industrial Electronics, vol. 50, No. 3 pp. 612-619, Jun. 2003.
Lee, et al. “Development of Indoor Navigation system for Humanoid Robot Using Multi-sensors Integration”, ION NTM, San Diego, CA pp. 798-805, Jan. 22-24, 2007.
Leonard, et al. “Mobile Robot Localization by tracking Geometric Beacons”, IEEE Transaction on Robotics and Automation, vol. 7, No. 3 pp. 376-382, Jun. 1991.
Li et al. “Robost Statistical Methods for Securing Wireless Localization in Sensor Networks”, Wireless Information Network Laboratory, Rutgers University.
Li et al. “Making a Local Map of Indoor Environments by Swiveling a Camera and a Sonar”, Proceedings of the 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 954-959, 1999.
Lin, et al.. “Mobile Robot Navigation Using Artificial Landmarks”, Journal of robotics System 14(2). pp. 93-106, 1997.
Linde “Dissertation, “On Aspects of Indoor Localization”” https://eldorado.tu-dortmund.de/handle/2003/22854, University of Dortmund, 138 pages, Aug. 28, 2006.
Luo et al., “Real-time Area-Covering Operations with Obstacle Avoidance for Cleaning Robots,” 2002, IEeE, p. 2359-2364.
Ma “Thesis: Documentation On Northstar”, California Institute of Technology, 14 pages, May 17, 2006.
Madsen, et al. “Optimal landmark selection for triangulation of robot position”, Journal of Robotics and Autonomous Systems vol. 13 pp. 277-292, 1998.
Martishevcky, “The Accuracy of point light target coordinate determination by dissectoral tracking system”, SPIE vol. 2591 pp. 25-30.
Matsutek Enterprises Co. Ltd “Automatic Rechargeable Vacuum Cleaner”, http://matsutek.manufacturer.globalsources.com/si/6008801427181/pdtl/Home-vacuum/10 . . . , Apr. 23, 2007.
McGillem, et al. “Infra-red Lacation System for Navigation and Autonomous Vehicles”, 1988 IEEE International Conference on Robotics and Automation, vol. 2, pp. 1236-1238, Apr. 24-29, 1988.
McGillem,et al. “A Beacon Navigation Method for Autonomous Vehicles”, IEEE Transactions on Vehicular Technology, vol. 38, No. 3, pp. 132-139, Aug. 1989.
Michelson “Autonomous Navigation”, 2000 Yearbook of Science & Technology, McGraw-Hill, New York, ISBN 0-07-052771-7, pp. 28-30, 1999.
Miro, et al. “Towards Vision Based Navigation in Large Indoor Environments”, Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, pp. 2096-2102, Oct. 9-15, 2006.
MobileMag “Samsung Unveils High-tech Robot Vacuum Cleaner”, http://www.mobilemag.com/content/100/102/C2261/, 4 pages, Mar. 18, 2005.
Monteiro, et al. “Visual Servoing for Fast Mobile Robot: Adaptive Estimation of Kinematic Parameters”, Proceedings of the IECON '93., International Conference on Industrial Electronics, Maui, HI, pp. 1588-1593, Nov. 15-19, 1993.
Moore, et al. A simple Map-bases Localization strategy using range measurements, SPIE vol. 5804 pp. 612-620, 2005.
Munich et al. “SIFT-ing Through Features with ViPR”, IEEE Robotics & Automation Magazine, pp. 72-77, Sep. 2006.
Munich et al. “ERSP: A Software Platform and Architecture for the Service Robotics Industry”, Intelligent Robots and Systems, 2005. (IROS 2005), pp. 460-467, Aug. 2-6, 2005.
Nam, et al. “Real-Time Dynamic Visual Tracking Using PSD Sensors and extended Trapezoidal Motion Planning”, Applied Intelligence 10, pp. 53-70, 1999.
Huntsberger et al. “CAMPOUT: A Control Architecture for Tightly Coupled Coordination of Multirobot Systems for Planetary Surface Exploration”, IEEE Transactions on Systems, Man, and Cybernetics—Part A: Systems and Humans, vol. 33, No. 5, pp. 550-559, Sep. 2003.
InMach “Intelligent Machines”, www.inmach.de/inside.html, 1 page, Nov. 19, 2008.
Innovation First “2004 EDU Robot Controller Reference Guide”, http://www.ifirobotics.com, 13 pgs., Mar. 1, 2004.
Wolf et al. “Robust Vision-based Localization for Mobile Robots Using an Image Retrieval System Based on Invariant Features”, Proceedings of the 2002 IEEE International Conference on Robotics & Automation, Washington, D.C. pp. 359-365, May 2002.
Wolf et al. “Robust Vision-Based Localization by Combining an Image-Retrieval System with Monte Carol Localization”, IEEE Transactions on Robotics, vol. 21, No. 2, pp. 208-216, Apr. 2005.
Wong “EIED Online >> Robot Business”, ED Online ID# 13114, 17 pages, Jul. 2006.
Yamamoto et al. “Optical Sensing for Robot Perception and Localization”, 2005 IEEE Workshop on Advanced Robotics and its Social Impacts, pp. 14-17, 2005.
Yata et al. “Wall Following Using Angle Information Measured by a Single Ultrasonic Transducer”, Proceedings of the 1998 IEEE, International Conference on Robotics & Automation, Leuven, Belgium, pp. 1590-1596, May 1998.
Yuta, et al. “Implementation of an Active Optical Range sensor Using laser Slit for In-Door Intelligent Mobile Robot”, IEE/RSJ International workshop on Intelligent Robots and systems (IROS 91) vol. 1, Osaka, Japan, pp. 415-420, Nov. 3-5, 1991.
Zha et al. “Mobile Robot Localization Using Incomplete Maps for Change Detection in a Dynamic Environment”, Advanced Intelligent Mechatronics '97. Final Program and Abstracts., IEEE/ASME International Conference, pp. 110, Jun. 16-20, 1997.
Zhang, et al. “A Novel Mobile Robot Localization Based on Vision”, SPIE vol. 6279, 6 pages, Jan. 29, 2007.
Euroflec Intellegente Monstre Mauele (English only except).
Roboking—not just a vacuum cleaner, a robot! Jan. 21, 2004, 5 pages.
SVET Computers—New Technologies—Robot vacuum Cleaner, 1 page.
Popco.net Make your digital life http://popco.net/zboard/view.php?id=tr—review&no=40 accessed Nov. 1, 2011.
Matsumara Camera Online Shop http://www.rakuten.co.jp/matsucame/587179/711512/ accessed Nov. 1, 2011.
Dyson's Robot Vacuum Cleaner—the DC06, May 2, 2004 http://www.gizmag.com/go/1282/ accessed Nov. 11, 2011.
Electrolux Trilobite, http://www.electrolux-ui.com:8080/2002%5C822%5C833102EN.pdf 10 pages.
Electrolux Trilobite, Time to enjoy life, 38 pages http://www.robocon.co.kr/trilobite/Presentation—Trilobite—Kor—030104.ppt accessed Dec. 22, 2011.
Facts on the Trilobite http://www.frc.ri.cmu.edu/˜hpm/talks/Extras/trilobite.desc.html 2 pages accessed Nov. 1, 2011.
Euroflex Jan. 1, 2006 http://www.euroflex.tv/novite—dett.php?id=15 1 page accessed Nov. 1, 2011.
Floorbotics, VR-8 Floor Cleaning Robot, Product Description for Manuafacturers, http://www.consensus.com.au/SoftwareAwards/CSAarchive/CSA2004/CSAart04/FloorBot/F.
Friendly Robotics, 18 pages http://www.robotsandrelax.com/PDFs/RV400Manual.pdf accessed Dec. 22, 2011.
It's eye, 2003 www.hitachi.co.jp/rd/pdf/topics/hitac2003—10.pdf 2 pages.
Robot Buying Guide, LG announces the first robotic vacuum cleaner for Korea, Apr. 21, 2003 http://robotbg.com/news/2003/04/22/lg—announces—the—first—robotic—vacu.
CleanMate 365, Intelligent Automatic Vacuum Cleaner, Model No. QQ-1, User Manual www.metapo.com/support/user—manual.pdf 11 pages.
Taipei Times, Robotic vacuum by Matsuhita about ot undergo testing, Mar. 26, 2002 http://www.taipeitimes.com/News/worldbiz/archives/2002/03/26/0000129338 accessed.
Tech-on! http://techon.nikkeibp.co.jp/members/01db/200203/10065011, 4 pages, accessed Nov. 1, 2011.
http://ascii.jp/elem/000/000/330/330024/.
IT media http://www.itmedia.co.jp/news/0111/16/robofesta—m.html accessed Nov. 1, 2011.
Yujin Robotics, an intelligent cleaning robot 'iclebo Q' AVING USA http://us.aving.net/news/view.php?articleId=7257, 8 pages accessed Nov. 4, 2011.
Special Reports, Vacuum Cleaner Robot Operated in Conjunction with 3G Celluar Phone vol. 59, No. 9 (2004) 3 pages http://www.toshiba.co.jp/tech/review/2004/09/59—0.
Toshiba Corporation 2003, http://warp.ndl.go.jp/info:ndljp/pid/258151/www.soumu.go.jp/joho—tsusin/policyreports/chousa/netrobot/pdf/030214—1—33—a.pdf 16 pages.
McLurkin “The Ants: A community of Microrobots”, Paper submitted for requirements of BSEE at MIT, May 12, 1995.
Grumet “Robots Clean House”, Popular Mechanics, Nov. 2003.
McLurkin, Stupid Robot Tricks: A Behavior-based Distributed Algorithm Library for Programming Swarms of Robots, Paper submitted for requirements of BSEE at MIT, May 2004.
Kurs et al, Wireless Power transfer via Strongly Coupled Magnetic Resonances, Downloaded from www.sciencemag.org, Aug. 17, 2007.
Nitu et al. “Optomechatronic System for Position Detection of a Mobile Mini-Robot”, IEEE Ttransactions on Industrial Electronics, vol. 52, No. 4, pp. 969-973, Aug. 2005.
On Robo “Robot Reviews Samsung Robot Vacuum (VC-RP30W)”, www.onrobo.com/reviews/AT—Home/vacuum—cleaners/on00vcrb30rosam/index.htm.. 2 pages, 2005.
OnRobo “Samsung Unveils Its Multifunction Robot Vacuum”, www.onrobo.com/enews/0210/samsung—vacuum.shtml, 3 pages, Mar. 18, 2005.
Pages et al. “Optimizing Plane-to-Plane Positioning Tasks by Image-Based Visual Servoing and Structured Light”, IEEE Transactions on Robotics, vol. 22, No. 5, pp. 1000-1010, Oct. 2006.
Pages et al. “A camera-projector system for robot positioning by visual servoing”, Proceedings of the 2006 Conference on Computer Vision and Pattern Recognition Workshop (CVPRW06), 8 pages, Jun. 17-22, 2006.
Pages, et al. “Robust decoupled visual servoing based on structured light”, 2005 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, pp. 2676-2681, 2005.
Park et al. “A Neural Network Based Real-Time Robot Tracking Controller Using Position Sensitive Detectors,” IEEE World Congress on Computational Intelligence., 1994 IEEE International Conference on Neutral Networks, Orlando, Florida pp. 2754-2758, Jun. 27-Jul. 2, 1994.
Park, et al. “Dynamic Visual Servo Control of Robot Manipulators using Neutral Networks”, The Korean Institute Telematics and Electronics, vol. 29-B, No. 10, pp. 771-779, Oct. 1992.
Paromtchik “Toward Optical Guidance of Mobile Robots”.
Paromtchik, et al. “Optical Guidance System for Multiple mobile Robots”, Proceedings 2001 ICRA. IEEE International Conference on Robotics and Automation, vol. 3, pp. 2935-2940 (May 21-26, 2001).
Penna, et al. “Models for Map Building and Navigation”, IEEE Transactions on Systems. Man. And Cybernetics. vol. 23 No. 5, pp. 1276-1301, Sep./Oct. 1993.
Pirjanian “Reliable Reaction”, Proceedings of the 1996 IEEE/SICE/RSJ International Conference on Multisensor Fusion and Integration for Intelligent Systems, pp. 158-165, 1996.
Pirjanian “Challenges for Standards for consumer Robotics”, IEEE Workshop on Advanced Robotics and its Social impacts, pp. 260-264, Jun. 12-15, 2005.
Pirjanian et al. “Distributed Control for a Modular, Reconfigurable Cliff Robot”, Proceedings of the 2002 IEEE International Conference on Robotics & Automation, Washington, D.C. pp. 4083-4088, May 2002.
Pirjanian et al. “Representation and Execution of Plan Sequences for Multi-Agent Systems”, Proceedings of the 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems, Maui, Hawaii, pp. 2117-2123, Oct. 29-Nov. 3, 2001.
Pirjanian et al. “Multi-Robot Target Acquisition using Multiple Objective Behavior Coordination”, Proceedings of the 2000 IEEE International Conference on Robotics & Automation, San Francisco, CA, pp. 2696-2702, Apr. 2000.
Pirjanian et al. “A decision-theoretic approach to fuzzy behavior coordination”, 1999 IEEE International Symposium on Computational Intelligence in Robotics and Automation, 1999. CIRA '99., Monterey, CA, pp. 101-106, Nov. 8-9, 1999.
Pirjanian et al. “Improving Task Reliability by Fusion of Redundant Homogeneous Modules Using Voting Schemes”, Proceedings of the 1997 IEEE International Conference on Robotics and Automation, Albuquerque, NM, pp. 425-430, Apr. 1997.
Prassler et al., “A Short History of Cleaning Robots”, Autonomous Robots 9, 211-226, 2000, 16 pages.
Radio Frequency Identification: Tracking ISS Consumables, Author Unknown, 41 pages, (NPL0127).
Remazeilles, et al. “Image based robot navigation in 3D environments”, Proc. of SPIE, vol. 6052, pp. 1-14, Dec. 6, 2005.
Rives, et al. “Visual servoing based on ellipse features”, SPIE vol. 2056 Intelligent Robots and Computer Vision pp. 356-367, 1993.
Robotics World Jan. 2001: “A Clean Sweep” (Jan. 2001).
Ronnback “On Methods for Assistive Mobile Robots”, http://www.openthesis.org/documents/methods-assistive-mobile-robots-595019.html, 218 pages, Jan. 1, 2006.
Roth-Tabak, et al. “Environment Model for mobile Robots Indoor Navigation”, SPIE vol. 1388 Mobile Robots pp. 453-463, 1990.
Sahin, et al. “Development of a Visual Object Localization Module for Mobile Robots”, 1999 Third European Workshop on Advanced Mobile Robots, (Eurobot '99), pp. 65-72, 1999.
Salomon, et al. “Low-Cost Optical Indoor Localization system for Mobile Objects without Image Processing”, IEEE Conference on Emerging Technologies and Factory Automation, 2006. (ETFA '06), pp. 629-632, Sep. 20-22, 2006.
Sato “Range Imaging Based on Moving Pattern Light and Spatio-Temporal Matched Filter”, Proceedings International Conference on Image Processing, vol. 1., Lausanne, Switzerland, pp. 33-36, Sep. 16-19, 1996.
Schenker, et al. “Lightweight rovers for Mars science exploration and sample return”, Intelligent Robots and Computer Vision XVI, SPIE Proc. 3208, pp. 24-36, 1997.
Shimoga et al. “Touch and Force Reflection for Telepresence Surgery”, Engineering in Medicine and Biology Society, 1994. Engineering Advances: New Opportunities for Biomedical Engineers. Proceedings of the 16th Annual International Conference of the IEEE, Baltimore, MD, pp. 1049-1050, 1994.
Sim, et al “Learning Visual Landmarks for Pose Estimation”, IEEE International Conference on Robotics and Automation, vol. 3, Detroit, MI, pp. 1972-1978, May 10-15, 1999.
Sobh et al. “Case Studies in Web-Controlled Devices and Remote Manipulation”, Automation Congress, 2002 Proceedings of the 5th Biannual World, pp. 435-440, Dec. 10, 2002.
Stella, et al. “Self-Location for Indoor Navigation of Autonomous Vehicles”, Part of the SPIE conference on Enhanced and Synthetic Vision SPIE vol. 3364 pp. 298-302, 1998.
Summet “Tracking Locations of Moving Hand-held Displays Using Projected Light”, Pervasive 2005, LNCS 3468 pp. 37-46 (2005).
Svedman et al. “Structure from Stereo Vision using Unsynchronized Cameras for Simultaneous Localization and Mapping”, 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 2993-2998, 2005.
Takio et al. “Real-Time Position and Pose Tracking Method of Moving Object Using Visual Servo System”, 47th IEEE International Symposium on Circuits and Systems, pp. 167-170, 2004.
Teller “Pervasive pose awareness for people, Objects and Robots”, http://www.ai.mit.edu/lab/dangerous-ideas/Spring2003/teller-pose.pdf, 6 pages, Apr. 30, 2003.
Terada et al. “An Acquisition of the Relation between Vision and Action using Self-Organizing Map and Reinforcement Learning”, 1998 Second International Conference on Knowledge-Based Intelligent Electronic Systems, Adelaide, Australiam pp. 429-434, Apr. 21-23, 1998.
The Sharper Image “e-Vac Robotic Vacuum, S1727 Instructions”www.sharperimage.com, 18 pages.
The Sharper Image “Robotic Vacuum Cleaner—Blue” www.Sharperimage.com, 2 pages, Mar. 18, 2005.
The Sharper Image “E Vac Robotic Vacuum”, www.sharperiamge.com/us/en/templates/products/pipmorework1printable.jhtml, 2 pages, Mar. 18, 2005.
TheRobotStore.com “Friendly Robotics Robotic Vacuum RV400—The Robot Store”, www.therobotstore.com/s.nl/sc.9/category.-109/it.A/id.43/.f, 1 page, Apr. 20, 2005.
TotalVac.com RC3000 RoboCleaner website Mar. 18, 2005.
Trebi-Ollennu et al. “Mars Rover Pair Cooperatively Transporting a Long Payload”, Proceedings of the 2002 IEEE International Conference on Robotics & Automation, Washington, D.C. pp. 3136-3141, May 2002.
Tribelhorn et al., “Evaluating the Roomba: A low-cost, ubiquitous platform for robotics research and education,” 2007, IEEE, p. 1393-1399.
Tse et al. “Design of a Navigation System for a Household Mobile Robot Using Neural Networks”, Department of Manufacturing Engg. & Engg. Management, City University of Hong Kong, pp. 2151-2156, 1998.
UAMA (Asia) Industrial Co., Ltd. “RobotFamily”, 2005.
Watanabe et al. “Position Estimation of Mobile Robots With Internal and External Sensors Using Uncertainty Evolution Technique”, 1990 IEEE International Conference on Robotics and Automation, Cincinnati, OH, pp. 2011-2016, May 13-18, 1990.
Watts “Robot, boldly goes where no man can”, The Times—pp. 20, Jan. 1985.
Andersen et al., “Landmark based navigation strategies”, SPIE Conference on Mobile Robots XIII, SPIE vol. 3525, pp. 170-181, Jan. 8, 1999.
Ascii, Mar. 25, 2002, http://ascii.jp/elem/000/000/330/330024/ accessed Nov. 1, 2011.
U.S. Appl. No. 60/605,066 as provided to WIPO in PCT/US2005/030422, corresponding to U.S. Appl. No. 11/574,290, U.S. publication 2008/0184518, filed Aug. 27, 2004.
U.S. Appl. No. 60/605,181 as provided to WIPO in PCT/US2005/030422, corresponding to U.S. Appl. No. 11/574,290, U.S. publication 2008/0184518, filed Aug. 27, 2004.
Derek Kurth, “Range-Only Robot Localization and SLAM with Radio”, http://www.ri.cmu.edu/pub—files/pub4/kurth—derek—2004—1/kurth—derek—2004—1.pdf. 60 pages, May, 2004, accessed Jul. 27, 2012.
Electrolux Trilobite, Jan. 12, 2001, http://www.electrolux-ui.com:8080/2002%5C822%5C833102EN.pdf Jul. 2, 2012, 10 pages.
Florbot GE Plastics, 1989-1990, 2 pages, available at http://www.fuseid.com/, accessed Sep. 27, 2012.
Gregg et al., “Autonomous Lawn Care Applications,” 2006 Florida Conference on Recent Advances in Robotics, Miami, Florida, May 25-26, 2006, Florida International University, 5 pages.
Hitachi ‘Feature’, http://kadenfan.hitachi.co.jp/robot/feature/feature.html, 1 page, Nov. 19, 2008.
Hitachi, http://www.hitachi.co.jp/New/cnews/hi—030529—hi—030529.pdf , 8 pages, May 29, 2003.
Home Robot—UBOT; Microbotusa.com, retrieved from the WWW at www.microrobotusa.com, accessed Dec. 2, 2008.
King and Weiman, “Helpmate™ Autonomous Mobile Robots Navigation Systems,” SPIE vol. 1388 Mobile Robots, pp. 190-198 (1990).
Li et al. “Robust Statistical Methods for Securing Wireless Localization in Sensor Networks,” Information Procesing in Sensor Networks, 2005, Fourth International Symposium on, pp. 91-98, Apr. 2005.
Martishevcky, “The Accuracy of point light target coordinate determination by dissectoral tracking system”, SPIE vol. 2591, pp. 25-30, Oct. 23, 2005.
Maschinemarkt Würzburg 105, Nr. 27, pp. 3, 30, Jul. 5, 1999.
Miwako Doi “Using the symbiosis of human and robots from approaching Research and Development Center,” Toshiba Corporation, 16 pages, available at http://warp.ndl.go.jp/info:ndljp/pid/258151/www.soumu.go.jp/joho—tsusin/policyreports/chousa/netrobot/pdf/030214—1—33—a.pdf, Feb. 26, 2003.
Paromtchik “Toward Optical Guidance of Mobile Robots,” Proceedings of the Fourth World Multiconference on Systemics, Cybermetics and Informatics, Orlando, FL, USA, Jul. 23, 2000, vol. IX, pp. 44-49, available at http://emotion.inrialpes.fr/˜paromt/infos/papers/paromtchik:asama:sci:2000.ps.gz, accessed Jul. 3, 2012.
Roboking—not just a vacuum cleaner, a robot!, Jan. 21, 2004, infocom.uz/2004/01/21/robokingne-prosto-pyilesos-a-robot/, accessed Oct. 10, 2011, 7 pages.
Sebastian Thrun, “Learning Occupancy Grid Maps With Forward Sensor Models,” Autonomous Robots 15, 111-127, Sep. 1, 2003.
SVET Computers—New Technologies—Robot Vacuum Cleaner, Oct. 1999, available at http://www.sk.rs/1999/10/sknt01.html, accessed Nov. 1, 2011.
Written Opinion of the International Searching Authority, PCT/US2004/001504, Aug. 20, 2012, 9 pages.
Denning Robscrub image (1989), 1 page.
Karcher Product Manual Download webpage: http://www.karcher.com/bta/download.en.shtml? ACTION=SELECTTEILENR&ID=rc3000&submitButtonName=Select+Product+Manual and associated .pdf file “5959-915en.pdf (4.7 MB) English/English” accessed Jan. 21, 2004 (16 pages).
Lumelsky, et al. “An Algorithm for Maze Searching with Azimuth Input”, 1994 IEEE International Conference on Robotics and Automation, San Diego, CA vol. 1, pp. 111-116, 1994.
Iirobotics.com “Samsung Unveils Its Multifunction Robot Vacuum”, www.iirobotics.com/webpages/hotstuff.php?ubre=111, 3 pages, Mar. 18, 2005.
Yun, et al. “Image-Based Absolute Positioning System for Mobile Robot Navigation”, IAPR International Workshops SSPR, Hong Kong, pp. 261-269, Aug. 17-19, 2006.
Yun, et al. “Robust Positioning a Mobile Robot with Active Beacon Sensors”, Lecture Notes in Computer Science, 2006, vol. 4251, pp. 890-897, 2006.
UBOT, cleaning robot capable of wiping with a wet duster, http://us.aving.net/news/view.php?articleId=23031, 4 pages accessed Nov. 1, 2011.
Kwon T-B et al: “Table Recognition through Range-based Candidate Generation and Vision-based Candidate Evaluation”. ICAR 2007, pp. 918-923.
Sadath M Malik et al. “Virtual Prototyping for Conceptual Design of a Tracked Mobile Robot”. Electrical and Computer Engineering, Canadian Conference on, IEEE, PI. May 1, 2006, pp. 2349-2352.
Related Publications (1)
Number Date Country
20090228165 A1 Sep 2009 US
Provisional Applications (1)
Number Date Country
60741442 Dec 2005 US
Continuations (1)
Number Date Country
Parent 11633869 Dec 2006 US
Child 12343430 US