This disclosure relates to a mobile autonomous service robot such as, e.g. a robot for processing surfaces (cleaning floors, mowing lawns, painting a surface, etc.). The disclosure relates, in particular, to methods for the remote control of such robots.
In recent years, autonomous mobile robots, in particular service robots, are being increasingly employed in the household, for example to clean floors or to monitor an apartment (night watchman function) or to transport objects into an apartment. Some of these robots can be monitored and controlled by means of a device that serves as a remote control (such as a computer). Also known are service robots that can send messages to a user or be remote controlled by a user over a long distance by employing an internet connection or mobile phone network to communicate with a mobile device of the user.
The known solutions either only provide for a locally employable control of the robot or for a control over the internet, whereas a control and monitoring of the robot over the internet is susceptible to hacking by unauthorized persons. Further, the robot may send information, e.g. concerning its status or the task that is being carried out or that is to be carried out by the robot, which is not necessarily of any aid to the user. For example, the information that the robot is incapable of cleaning a room due to a closed door may be of only minor relevance for the user if he/she is not in the proximity of the apartment, as in this case he/she cannot immediately remedy the problem.
This application discloses a method for monitoring and controlling an autonomous mobile service robot by means of a mobile device, without the user having to devote undue attention to the communication channel between the device and the robot. Further, the “intelligence” of the robot's communication with the user should be improved.
A system having at least one mobile autonomous robot (service robot) and at least one mobile device for the control the robot is described herein. In accordance with one embodiment, the system further comprises at least one communication server that has a network interface with access to the public internet. The robot and the mobile device are configured to communicate both via a first connection, as well as via at least one second connection, whereas the second connection is a local wireless communication connection and the first connection is an indirect communication connection, in which the communication server relays the communication between the robot and the mobile device. If one of the two connections is interrupted or setting up the respective communication connection fails, an attempt is made to change the communication connection.
Further, a method for communicating between a mobile device and a mobile autonomous robot is described herein. In accordance with one embodiment, the method includes setting up a communication connection between the robot and the mobile device via a first connection or via a second connection, whereas the second connection is a local wireless communication connection and the first connection is an indirect communication connection, in which a communication server that is accessible via the internet relays the communication between the robot and the mobile device. The method further includes changing the communication connection from the second connection to the first connection if the second connection is interrupted or setting up the second connection fails.
Further, a method for coupling a device with a mobile autonomous robot is described herein. In accordance with one embodiment, first the robot initiates a communication connection with a server and the mobile device sends a request to the server. The server informs the robot of the request from the device by means of a message. The robot then waits for a certain user interaction, by means of which the request is confirmed by the user. If the user interaction is carried out within a specified period of time, the robot will carry out control commands that are henceforth sent to the robot by the device via the server, otherwise a control of the robot via the communication connection with the server will be denied.
Further, a system with at least one autonomous mobile robot and at least one mobile device is described herein. In accordance with one embodiment, the robot and the device are configured to communicate via at least one first wireless connection. The robot is configured to determine information concerning the location of a user (e.g. based on the location of the device) and, dependent on the determined information concerning the location, to decide whether, and if so, then which information is to be sent to the device and/or whether, and if so, then which tasks are to be carried out, interrupted, continued or ended by the robot.
Further, a system with at least one autonomous mobile robot and at least two mobile devices is described herein. In accordance with one embodiment, each of the mobile devices is configured to communicate with the robot. The robot is configured to allow changes to the information stored in the robot by one of the mobile devices and to send information concerning possible changes to at least one other mobile device.
Further, a system with at least one autonomous mobile robot, at least two mobile devices and at least one server with internet access is described herein. Each of the mobile devices is configured to communication with the robot, whereas the server relays the information. The robot and server are configured to synchronize the information stored in the robot with the information stored in the server.
This disclosure also relates to an autonomous mobile robot. In accordance with one embodiment, the robot comprises a navigation module that is configured to navigate the robot through the area of robot employment with the aid of a map of the area of employment and sensor information available in the robot. A detection unit of the robot is configured to detect at least one of the following events: the user enters or leaves the area of robot employment, has entered or has left the area or is just in the process of entering or leaving the area. A control module of the robot is configured to trigger a previously specified action of the robot when the detection unit detects one of the events, whereas the previously specified action consists in the robot moving to a location previously specified on the map and beginning a previously specified operation.
Further, a system with a first and with at least one second autonomous mobile robot is described herein. In accordance with one embodiment, these are configured to communicate with each other, whereas the first robot is configured to carry out a first operation and the second robot is configured to carry out a second operation. The first robot is configured to receive a command from a user to begin an operation, to interrupt and/or terminate it, whereas the first robot informs the second robot of the user's command and the second robot is configured, based on this information, to decide whether to begin the second operation, to begin the operation with a delay or to interrupt and/or terminate it.
In accordance with a further embodiment of the aforementioned system, at least the first robot is configured to detect whether a user leaves or enters the area of robot employment, whereas the first robot informs the second robot when it detects that a user leaves or enters the area of robot employment and the second robot is configured, based on this information, to decide whether to begin the second operation, to begin it with a delay or to interrupt and/or terminate it.
Further embodiments refer to a method for controlling autonomous mobile robots with at least one mobile device. It should be explicitly mentioned that the various features described in the context of an embodiment may also be used (provided they are not explicitly otherwise described) with all other embodiments in order to thus create further embodiments.
Various embodiments are described in the following based on the examples illustrated in the figures. The figures are not necessarily true to scale and the embodiments are not limited to only the shown aspects. Instead importance is placed on illustrating the underlying principles. The figures show:
In accordance with the embodiments described herein, the user of an autonomous mobile service robot is to be enabled to control and monitor the service robot with the aid, for example, of a smartwatch, a smartphone or a tablet PC. This is to be possible both locally in the area of robot employment (e.g. the apartment of the user), as well as from a distance (outside of the area of robot employment). This should bring about no significant changes for the user in the handling of a mobile device for the control and monitoring of the service robot.
In order to allow for a secure and reliable communication between the robot and a mobile device located outside of the area of robot employment, both the robot and the mobile device can connect with a communication server that is accessible via the internet. In the area of robot employment (such as, for example, in the apartment of the user), however, the control of the robot should not take place via the internet, but instead via a direct connection between the robot and the mobile device. This makes it possible to avoid an unnecessary (and possibly fee based) data transmission (on the user and the server side) and should improve the quality and the security of the data exchange. The direct connection between the robot and the mobile device may be set up, e.g. via a home network such as, for example, via a WLAN (Wireless Local Area Network). Further details concerning possible types of connections will be explained further below. A server is not necessarily understood, in this context, to be a specific computer or specific hardware, but rather a server service (server daemon) that is executed on a computer. The communication server is thus software that transmits the communication between the device and the robot.
The autonomous mobile robot can carry out, automatically as a service robot, one or more tasks such as, e.g., the cleaning or monitoring of an apartment or of a building (or of a part thereof), as well as the transport of objects within a building. Each robot may have a specific identification number (UID, Unique Identification Number) with which, for example, it can log in to the mentioned communication server. Further, the user may also give a name to his/her robot. In order to be able to autonomously carry out a task in an area of employment, the robot generally has devices for navigation (cf., for example,
The mobile device may be, e.g. a smartwatch, a smartphone, a tablet PC or a portable computer having a suitable application (abbreviated: app) that provides a graphical user interface. The autonomous mobile robot and the mobile device each have at least one communication interface. This may be, for example, a WLAN module, with which a connection with a local WLAN (in the area of robot employment) can be set up. The WLAN can facilitate a connection to the Internet and/or a direct connection between the robot and the mobile device. Other known possible communication interfaces include wireless connections such as, for example, the mobile telephone network (UMTS, LTE), Bluetooth, ZigBee, Z-Wave or Near Field Communication (NFC), or direct connections such as, for example, a USB interface.
In accordance with the embodiment described herein, the control and monitoring of the robot with the aid of the mobile device may encompass various aspects. The mobile device may be employed, for example, to display status information, to make use of calendar functions and to enter commands for the robot. Status information includes, for example, the display of the charging status, a map of the area of robot employment with information regarding a cleaning or a monitoring of the area or the location of the robot in its area of employment. With the aid of the mentioned calendar function, regularly occurring tasks or one-time tasks that lie in the future can be assigned to the robot. This means that a schedule can be drawn up for the tasks to be carried out by the robot. Examples of this include “Begin a cleaning every week on Monday at 9 am” or “Monitor the apartment this evening from 8 to 10 pm to detect unusual activities”. Commands given to the robot include, for example, the start of a new task or the interruption of a task being carried out by the robot. This may be, e.g., the cleaning or the monitoring of a specifiable area or movement to a specifiable location such as, for example, the base station of the robot.
In accordance with the embodiments described here, the mobile device for the control and monitoring of the robot can send various requests to the robot which, in turn, sends a corresponding response back to the mobile device and/or carries out a corresponding action. These requests can be divided into two groups:
The above mentioned communication server can provide a communication service (server service) that allows for the communication between the robot and the mobile device over great distances and that is permanently accessible, e.g. via the internet. For this purpose, both the robot and the mobile device can register with (log in to) the communication server. The communication server is configured, for example, as a cloud service, as a virtual server, as an individual service computer (host) or as a network of service computers (cluster). In some embodiments, the connection between the robot and the communication server or between the mobile device and the communication server cannot take place directly, but only through relay over a further server service. For example, for home automation (smart home), the control of numerous different devices can be bundled for the mobile device into one application. This application then contacts a central server that relays or forwards the communication to the particular communication server that is specialized for each device (for example, the service robot in question).
Various possibilities for setting up a connection between a mobile autonomous robot 1 and a mobile device 2 are shown in
In the example of
In the example in accordance with
In the example shown in
In accordance with one embodiment, the mobile device 2 can send a request via the communication server 3 to the robot 1, asking which local connection possibility it can currently enable. The robot 1 responds to the communication server by informing it which WLAN (SSID) it is logged in to and with which IP address. The mobile device can then check whether this WLAN is available from its present location and can set up a local connection. This procedure is advantageous when, for example, for technical reasons an mDNS request is not possible or is too slow within a given WLAN. Furthermore, the device 2 can observe over time whether it has entered the coverage area of the WLAN and can then directly connect with the robot via WLAN without making a repeated mDNS request.
In accordance with one embodiment, when the application on the device is used, the mobile device 2 can monitor the types of connection available to the robot 1. If the current connection (for whatever reason) is disrupted, it will attempt to set up a new connection via an alternative path. The device can also search for alternative connection possibilities regardless of whether the current connection between the mobile device 2 and the robot 1 is disrupted. When the current connection uses the communication server 3, the mobile device 2 can constantly or periodically search for a local connection possibility and can use the latter instead of the connection via the communication server 3. The mobile device 2 can inform the user and/or the robot 1 of a change in the connection possibility and enable them to react to this. Alternatively or additionally the robot can be configured to detect a change in the communication connection to the device 2 and, if necessary, react to a detected change. For example, the robot can determine whether a command or a request has been received directly by the device 2 or by a communication server (e.g. using the IP address).
In one example, the robot 1 and the mobile device 2 are logged in to the same (local) WLAN and communicate via this connection. If the user with the mobile device 2 leaves the apartment, the mobile device 2 loses the connection to the local WLAN in the area of robot employment and thus also loses the direct connection with the robot. The mobile device 2 can now attempt to set up a connection to the communication server 3 (for example, via the mobile phone network), in order to connect by these means with the robot 1. At the same time, the user, for example, is asked whether the robot 1 should begin a task, for example, the monitoring or cleaning of an apartment. Alternatively, the robot can be informed of the change in the connection type. It can then test whether it should change its current status by, for example, automatically beginning a task. For example, the user leaving the apartment can automatically trigger the robot 1 to monitor or clean the apartment.
In a further example, the user with the mobile device 2 comes home. In this case the mobile device 2 detects the new local connection possibility to the robot 1, for example via a mutually shared WLAN or a Bluetooth radio connection. Now the user can be given a brief status report. In particular, the user can be informed of particular events. For example, the user can be informed that the robot 1 was not able to clean a room because a door was closed. Since the user is now at home he can open the door so that the robot 1 can complete the cleaning. Alternatively, the robot 1 can react to the possibility of a new local connection, for example, by ending its currently performed task such as the cleaning or monitoring of an apartment. Thus it can be avoided that the user is disturbed by the work of the robot 1. The possibility of sending messages to the user when the type of connection changes or dependent on the type of connection will be explained in greater detail further below.
In order to allow for the type of connection between mobile device 2 and robot 1 by means of the communication server 3, as shown in
A further example is shown in
In a simple example, the communication server 3 transmits the communication between robot 1 and mobile device 2 directly. To do so, however, the connection between robot 1 and communication server 3, as well as that between the mobile device 2 and the communication server 3, must be permanent and very good. In order to be able to quickly respond to requests from the mobile device 2 regarding the status of the robot 1, it is advantageous for the communication server 3 to regularly request and save information concerning the current status of the robot 1. When the mobile device 2 then sends a request for information only, the communication server can thus immediately send a response, without having to first forward the request to the robot and to wait for its response after having received the server's request and before responding to it. When doing so it can also be tested how much time has elapsed since the last connection to the robot 1, i.e. to what extent the information received and saved by the robot 1 is up to date. If the elapsed amount of time exceeds a specified value (for example, five minutes), then the mobile device 2 is also informed of the fact that there is currently no connection to the robot. The communication server will then intermediately save any command given to the robot 1 until the robot 1 has set up a new connection. For this, a specified time (for example, 10 seconds) can be predetermined, after which the mobile device is informed of the fact that the robot is not accessible. For example, the user can decide (e.g. by entering this into the device 2), whether a command sent to the communication server 3 should expire after the elapsed amount of time or whether it should remain in effect.
In order to reduce the amount of data transmitted between robot 1 and communication server 3, the information requested by the communication server 3 can, for example, be dependent on the last known status of the robot. For example, no update of a map need be requested when the robot 1 is recharging itself at its base station.
In accordance with the embodiment outlined in
An autonomous mobile service robot 1 gathers, by means of its numerous sensors, sensitive data about its area of employment, including, for example, private data involving the user's apartment. Such data should be protected from unauthorized access by third parties by means of methods such as, for example, encryption using TLS (Transport Layer Security) certificates. At the same time, however, the operability of the robot 1 should not be impaired by too many password requests.
In order to introduce an additional level of security that will impede access of unauthorized parties to sensitive data, in one exemplary embodiment the user may provide proof of his access authorization when setting up the robot control via the communication server 3. This may be carried out, as shown, e.g., in
This method for setting up the robot control by means of the communication server 3 comprises, in one particular embodiment, the following five steps:
In the following, the described concept is explained in greater detail by means of an example. In step (a) the user enables the robot 1 to set up a connection to the communication server 3. It does so, for example, by giving the robot 1 access to a WLAN router 4 that is connected with the internet. The robot 1 logs in to the communication server 3, for example using its UID (Unique Identification Number), allowing it to be clearly identified.
In step (b) the mobile device 2 connects with the communication server 3 and informs the latter that a connection to the robot 1 is desired. This is carried out, for example, by the user providing an identifier of the robot (e.g. the UID of the robot 1) and a corresponding password. The user can find this information, for example, in the robot's 1 documentation (e.g. user manual). The UID or another identifier of the robot (e.g. a digital certificate) may also be transmitted from the robot 1 to the mobile device 2, for example, via a direct connection (e.g. by means of a point-to-point radio connection such as, e.g. via a Bluetooth connection or by means of NFC or via a WLAN). An alternative “transmission path” between robot 1 and device 2 may comprise the robot 1 having a 2D barcode such as, e.g. a QR code (Quick Response Code), located, for example, on its underside and in the barcode being scanned with the device 2. The robot 1 may also generate a (pseudo) random identifier. If the identifier is to be transmitted via a QR code, then this may be shown, for example, in a display on the robot. In addition, it is possible for the user to log in to a user account on the communication server 3.
In step (c) the communication server 3 informs the robot 1 that a mobile device 2 requests a connection with the robot 1 via the communication server 3. After this, in the fourth step (d), the robot 1 waits for a direct user interaction to take place within a specified period of time. In step (d) there are numerous possibilities for a direct user interaction with the robot 1 that are required in order for the user to physically assume control of the robot. For example, the user interaction may be carried out tactilely by pressing a button on the robot 1 or by triggering a touch-sensitive sensor (e.g. bumper switch) of the robot 1. Further possibilities include the deliberate triggering of a different sensor (such as, for example, an accelerometer, a gyroscope, a wheel-contact switch, a floor sensor, a crash sensor or sensors for measuring the distance to obstacles) by the user by, for example, lifting, bumping, shifting or rotating the robot. More detailed instructions for the user on how to do this are displayed, for example, on the mobile device. In addition, for example, the button to be pressed may light up or blink. If actuation by the user takes place within a specified period of time by means of a physical interaction between user and robot that is detectable by the robot (e.g. the pressing of a button, lifting or turning the robot, etc.), then, in accordance with step (e), the robot 1 sends a confirmation to the communication server 3. In the opposing case, control over the robot via the communication server by means of the mobile device will be denied. After the connection has been confirmed, the authorization of the user and/or the mobile device 2 to control the robot 1 via the communication server 3 can be recorded in a data base on the communication server 3. As an alternative, for example, a digital certificate is saved on the mobile device 2, by means of which it can prove its authorization to control the robot 1 to the communication server 3.
The interaction between user and robot may also consist in the user scanning a code (e.g. a barcode or a QR code) found on the robot or on a display of the robot using a camera integrated in the mobile device. The scanned code is then transmitted via the server 3 to the robot 1, which can verify the code (for this purpose the code is also stored in a non-volatile memory in the robot). The code may be, for example, an identification number that distinguishes the robot (UID, e.g. consisting of its model and series numbers) or a (pseudo) random code or a digital certificate. In this case the robot must show the QR code in a display on the robot. This code, logically, differs from the identifier described above in step (b) that is transmitted to the server. Instead of a code, at least one checksum or a hash value may also be scanned and transmitted.
A further possible direct user interaction in correspondence with step (d) would be a short range communication connection between robot 1 and mobile device 2. Various methods may be used for this, for example, a wire-based connection, for example, by means of USB, or a wireless connection, for example, by means of Bluetooth, NFC or infrared can be used. In this case, for example, the communication server 3 may send to the mobile device 2 via the robot a digital certificate as proof of authorization to control the robot. For this the certificate must be transmitted to the robot, e.g. by means of an infrared or Bluetooth data connection.
The procedure just described should preferably be carried out for every robot 1 and for every user and/or every mobile device 2. In the case of multiple users and/or mobile devices, there may be a master user and/or a master device that maintains and assigns the authorization to monitor and control the robot for other authorized users and/or devices. In this manner one user, for example, can have a cleaning map of the robot shown to him while a second user additionally has the authorization to have currently recorded monitoring (surveillance) images shown to him.
One of the desired automatic operations that an autonomous mobile robot 1 should perform is to inform the user of the results of the tasks carried out by the robot and of problems that might have occurred while doing so. For example, the robot may inform the user that a room cannot be cleaned because of a closed door. This information is only relevant for the user, however, if he is capable of changing these circumstances. If the user is away from home at that point in time (meaning he is not in the proximity of the area of robot employment) and receives such a message (“door closed”), then the user cannot in any way change the situation. The message may then even present a disturbance or the user may forget the message by the time he arrives home. Whether and when sending a piece of information to the user makes sense may thus also depend on the location of the user.
In accordance with one embodiment, the relevance of messages is determined according to whether the user is at home (i.e. in or in the vicinity of the area of robot employment), out of home or currently returning home. When the user is at home, he may be interested in receiving messages if the robot is about to begin an activity that might potentially disturb the user. Such an activity would be, for example, the cleaning of a room. A further interesting message would be that the robot has detected a problem (e.g. a closed door) that it cannot remedy on its own. When the user is away from home, security-related messages from a monitoring robot or those concerning serious problems will be of interest (for example, that the robot cannot begin a planned cleaning because it is locked in a room or located on the wrong floor). If the service robot is carrying out monitoring tasks (night watchman function), certain events could trigger different alarms, depending, e.g., on whether the user is present. Upon returning home after having been out, he might be interested in what state the apartment is in (for example, the results of the cleaning) and in messages concerning problems that the robot could not independently remedy in the user's absence.
The messages sent to the user by the robot may concern its own field of operation, but may also extend beyond it. Two examples are shown in
If numerous users are to be able to control the same robot, the robot can check which of the users are present and send a corresponding message to the appropriate user(s). For example, in one household there may be two people who are authorized to control the robot. If the robot 1 detects a problem that it cannot remedy on its own, such as a closed door, then it checks which of the two people is at home and pointedly informs this person by sending a message to the person's device.
The types of connection described above (see
In some cases it may be useful to evaluate additional information concerning the location of the user. An example of this is shown in
In addition to evaluating the connection types, there are further possibilities for determining the presence of a user or of another person (e.g. a child, an unauthorized person, etc.). For this, for example, the robot may use sensors, for example, a camera or a microphone, to detect the movement of a person. In the case of camera-based detection, e.g. algorithms for face recognition may be employed. Further, for devices that the user does not usually carry with him but that are used to control and monitor the robot (for example, (smart) television, PC, possibly also a tablet PC), the usage activity can be tested and on this basis it can be decided whether a message will be forwarded to the mobile device 2. The presence of a user may also be detected with the aid of other devices, such as an alarm system. Thanks to the increasing interconnectedness of a wide variety of household appliances (IoT, Internet of Things), the robot can also make use of usage data from another robot, a coffee machine, a stove, electric blinds or even those of fitness and health trackers. Thus, a specific command given to a device by a user may indicate that he intends to leave the home. For example, he may request a transport robot to bring him his street shoes. The transport robot conveys this command to a second robot that is used to monitor and/or clean the house, triggering the latter to automatically carry out one of the above described searches for forgotten items, open windows, etc. A cleaning robot can also check whether all areas to be cleaned are accessible and, if necessary, call to the user's attention a closed door before the user leaves the house.
The mobile robot 1 can easily send a message to the user when a connection to the mobile device 2 has been set up. If there is no direct wireless connection, the robot may revert to other notification services such as, for example, email, SMS or a push service such as, for example, the Apple Push Notification Service or the Google Cloud Messaging service. These means of messaging can be initiated by the robot 1 or, alternatively, by the communication server 3.
For example, the robot may first attempt to send a message to the mobile device over a direct connection. If this does not work, then the message can be sent to the communication server. The communication server may again attempt to send the message directly to the mobile device. For example, the server may wait for a specified amount of time for the mobile device to contact it. After this time has expired, the communication server may send, for example, a push message or an SMS.
In accordance with one embodiment, after the user has left the house the robot can specifically check whether the windows and house doors are closed and/or whether the electric appliances have been turned off and can either close the former or turn off the latter on its own and/or inform the user of a critical situation (as described in the passage above). When the user arrives home the robot may come into the corridor to greet him, entertain him or bring him his house slippers. After the user has left the corridor the robot may begin to clean it in order to remove any dirt brought in with the user's street shoes. Before doing so the robot either attempts to detect whether the user has left the corridor, or it waits a specified amount of time. A further possibility is a transport robot that offers to carry items that the user might have purchased into, for example, the kitchen.
In order to realize all these examples, it can be advantageous for the autonomous mobile robot to possess a map of its area of employment so that it can easily navigate the area of employment. Certain positions or subareas of the area of robot employment may be linked to specific actions that can be triggered when the user reaches or leaves home. Linking a given area to a corresponding action can be carried out by the user. In order to improve the utility of the robot even further, it can be configured to recognize and save the relevant areas automatically. Thus, for entertainment purposes, the user can adjust the robot to greet him when he arrives home (linking “corridor” to the action “greeting” and the event “user arrives home”), and/or to offer a task such as the transport of purchased items. The areas that are relevant to this task can be learned by the robot or interpreted based on specifiable criteria from a map.
For example, the robot can sector the map of its area of employment into subareas that correspond, for example, to the rooms of an apartment. The user may give these subareas designations (tags) such as, for example, “living room”, “kitchen” or “corridor”. Other methods are also known for the robot to recognize the function of a given area (e.g. identifying typical objects by means of image processing). In order to greet the user when he arrives home, the robot can thus enter the subarea designated as “corridor”.
Alternatively, the robot can learn where the user's entrance and exit areas are located in its area of employment. During its first exploratory run through the area of employment (during which the map is created) the robot can, for example, detect (for example, based on information from the mobile device) that the user is arriving home. The robot will then look for the user within its area of employment and the location at which the user is finally found will be saved. During a subsequent run, the robot can first move to the saved location, enabling it to find the user much more quickly. The location at which the robot discovers the user will again be saved. After a few repetitions, the thus determined and saved location can be reasonably described as the entrance and exit area and this information can be shared with other robots.
When the robot detects that a user has arrived at or left the area of robot employment, it may happen that it is, at that moment, occupied with a different activity. In this case the robot must check whether it should interrupt the activity in order to carry out a new one. For example, the robot may have been assigned the task of cleaning the corridor after the user has arrived home. If at that moment the robot is recharging its battery, it may make more sense for the robot to postpone cleaning the corridor until the battery is fully charged. In other cases, postponing the activity may not be appropriate, in which case the activity may be omitted altogether, for example, if the robot is located on a different floor of the building and cannot access the corridor. Further, the robot may begin, postpone or omit an activity depending on whether further users of the robot are present in the home or not.
Robots exist that, for example, store their map data on external computers or in a cloud service. If the connection to the computer is interrupted, however, they are no longer able to carry out their tasks. In order to ensure that the robot is able to operate autonomously, all data that the robot requires to carry out its assigned task, such as map data for navigation, but also data for connecting to local networks, are stored in the robot. In a similar manner, user specific data may also be stored in the robot (e.g. in a memory, cf.
Further, it may also occur that numerous users employ the same robot. In this case it may be advantageous to install a user administration in the robot that is capable of handling various devices and users. For example, the user administration can carry out a prioritization of users and/or devices if they all want to perform changes on the robot at the same time. For example, a first user might command the robot, via the communication server, to carry out a cleaning. A second user who, at this time, is at home might be disturbed by this and may therefore, via a local connection, command that the robot return to its base station. In this case, the command that is transmitted via the local network will have a higher priority and will therefore be carried out. The first user, if necessary, will be informed of this.
Further, various authorizations may be granted to the users. Thus, for example, one user may be allowed to assign tasks to the robot in a schedule (calendar function), while a second is only permitted to view the planned time periods. Still further, for each calendar entry, the user who made the entry can also be saved. If the entry is changed by a further user, the first user will be informed of this. If needed he may then either confirm or rescind the change.
In the case of various users and/or devices, the robot can be configured as to which user and/or device will be informed of a specific event. A further possible configuration is the choice of the mobile device that can be used to determine the location of a user in order to send the location-dependent (depending on the current location of the user) messages described above. For example, a smartwatch or a smartphone that the user usually carries with him will be used, whereas a tablet PC that is usually left at home will be ignored.
When numerous robots are employed in a household, the two or more robots inform each other of the personalized adjustments made by the user. Thus, the user must only once personalize the map of a first robot by designating areas and subareas. Afterwards, the first robot informs a second robot of this personalized map data. This takes place, for example, via a direct wireless connection such as Bluetooth or via a local network such as, for example, a WLAN. The exchange of map data among the robots allows for a general improvement of the map data available to a robot. For example, a vacuuming robot and a mopping robot may be employed in one apartment, whereas the mopping robot, for example, will not move onto a carpet. By exchanging map data, the mopping robot obtains additional information about its environment. In addition to this, it can show to the user the same map as that of the vacuuming robot. Further, the vacuuming robot can inform the mopping robot of a command from the user to clean the apartment and may additionally let it know which surfaces it has already vacuumed.
At the same time, the robots can thus use each other as backups, making it easier to replace a defect robot. In one example one robot is employed on each floor of a house with two floors. Exchanging map data allows, for example, the robot from the first floor to be directly employed on the second floor without the need for an additional exploratory run. For example, a robot newly purchased by the user can obtain the map and other data from a robot already employed in the household, which significantly reduces the effort needed to set up a robot. For example, in this manner an exploratory run, during which the robot becomes acquainted with its area of employment and compiles a map thereof, can be omitted.
If a communication server is used, it may naturally also be utilized to store additional data that is not immediately needed for a robot to carry out a task. For example, the robot may maintain data concerning the progression of its activity for a specified period of time. Due to the limited memory capacity of the robot, this time period is also very limited. In order to extend this period of time or permanently save individual events, the communication server or a cloud memory may be used. In addition to this, the server may also be used as a backup for the robot, which, in particular when only one robot is employed, affords the advantages of a backup described above.
In the present example, the robot 1 includes a drive module 101 that, for example, includes electric motors, gears and wheels. It may further include a power source and power electronics used to drive the robot's drive train. The drive module 101 enables the robot to—theoretically—reach any point within the area of robot employment. As mentioned, robot 1 may have access to a communication link 104 via communication module 103 that may include an interface to a local computer network (WiFi, ZigBee, etc.), the internet or to a point-to point wireless link (e.g. Bluetooth) or any other suitable communication means. In the depicted example, the robot 1 is registered with a wireless access point included in the home router 4. Further, the processors and memories may be utilized to optimize the usage of computing resources, power resources and/or connectivity while executing the operations conducted by the robot. As a result, such features provide substantial operational efficiencies and improvements over existing technologies.
As many common service robots, the robot 1 may further include a processing module 102 that may be, for example, a cleaning module configured to clean the floor. The cleaning module my include brushes, a vacuum unit or the like. In order to be able to perform a task autonomously, the robot 1 may have a navigation module 110, with which the robot may orient itself and navigate across the robot deployment area using so-called navigation features, i.e. features, with which the robot 1 may orient itself in its environment, such as landmarks (furniture, doorways, corners or a room, walls, etc) that may be detected by the robot 1 using its sensor module 130. The navigation module 110 may, for example, employ an obstacle avoidance strategy and/or a SLAM (simultaneous localization and mapping) algorithm in connection with one or more electronic maps of the robot deployment area. Various suitable algorithms are as such known and thus not further discussed herein. The map(s) of the area of robot employment may be newly generated by the robot during a deployment (e.g. while performing a task). Alternatively, a previously stored map may be used during a deployment. The stored map may have been generated by the robot itself during a preceding deployment or provided by the user or by another robot. The memory module 120 may include a non-volatile memory (e.g. a solid state disc, SSD) and may contain the maps of the area of robot employment. Alternatively, the maps may be stored externally, e.g. in or by the external data processing device (e.g. in a computer located in the apartment or by a cloud server such as the server 3).
The sensor module 130 may include one or more sensors for measuring distances to objects in the robot's environment such as optical sensors, acoustic sensors which operate using as such known triangulation or time-of-flight measurement (e.g. triangulation sensor, time-of-flight camera, laser range finder, ultrasonic sensors, etc.). Other suitable sensors may be cameras (in connection with image processing techniques), tactile sensors, gyroscopic sensors, inertial measurement units (IMUs), odometers and/or floor clearance sensors. Various suitable sensors are as such known and used in currently available autonomous robots and thus not further discussed herein.
While the machine-readable medium is shown in an example embodiment to be a single medium of memory 152, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions. The term “machine-readable medium” shall also be taken to include any medium that is capable of storing, encoding or carrying a set of instructions for execution by the machine and that causes the machine to perform any one or more of the methodologies of the present disclosure.
The term “machine-readable medium” shall accordingly be taken to include, but not be limited to: memory devices, solid-state memories such as a memory card or other package that houses one or more read-only (non-volatile) memories, random access memories, or other re-writable (volatile) memories; magneto-optical or optical medium such as a disk or tape; or other self-contained information archive or set of archives is considered a distribution medium equivalent to a tangible storage medium. The “machine-readable medium” may be non-transitory, and, in certain embodiments, may not include a wave or signal per se. Accordingly, the disclosure is considered to include any one or more of a machine-readable medium or a distribution medium, as listed herein and including art-recognized equivalents and successor media, in which the software implementations herein are stored.
Although various embodiments have been illustrated and described with respect to one or more specific implementations, alterations and/or modifications may be made to the illustrated examples without departing from the spirit and scope of the appended claims. With particular regard to the various functions performed by the above described components or structures (units, assemblies, devices, circuits, systems, etc.), the terms (including a reference to a “means”) used to describe such components are intended to correspond—unless otherwise indicated—to any component or structure that performs the specified function of the described component (e.g., that is functionally equivalent), even if it is not structurally equivalent to the disclosed structure that performs the function in the herein illustrated exemplary implementations of the present application.
Further, the purpose of the Abstract of the Disclosure is to enable the U.S. Patent and Trademark Office and the public generally, and especially the scientists, engineers and practitioners in the art who are not familiar with patent or legal terms or phraseology, to determine quickly from a cursory inspection the nature of the technical disclosure of the application. The Abstract of the Disclosure is not intended to be limiting as to the scope in any way.
Finally, it is the applicant's intent that only claims that include the express language “means for” or “step for” be interpreted under 35 U.S.C. 112. Claims that do not expressly include the phrase “means for” or “step for” are not to be interpreted under 35 U.S.C. 112.
Number | Date | Country | Kind |
---|---|---|---|
10 2015 121 666.3 | Dec 2015 | DE | national |
This application is a Continuation-In-Part Application and claims the benefit of PCT/EP2016/080235 designating the United States, filed Dec. 8, 2016, the entirety of which is herein incorporated by reference and which claims priority to German Patent Application No. DE 10 2015 121 666.3, filed Dec. 11, 2015.
Number | Name | Date | Kind |
---|---|---|---|
4674048 | Okumura | Jun 1987 | A |
4740676 | Satoh et al. | Apr 1988 | A |
4777416 | George, II et al. | Oct 1988 | A |
5109566 | Kobayashi et al. | May 1992 | A |
5260710 | Omamyuda et al. | Nov 1993 | A |
5284522 | Kobayashi et al. | Feb 1994 | A |
5377106 | Drunk et al. | Dec 1994 | A |
5402051 | Fujiwara et al. | Mar 1995 | A |
5696675 | Nakamura et al. | Dec 1997 | A |
5787545 | Colens | Aug 1998 | A |
5995884 | Allen et al. | Nov 1999 | A |
6389329 | Colens | May 2002 | B1 |
6532404 | Colens | Mar 2003 | B2 |
6594844 | Jones | Jul 2003 | B2 |
6605156 | Clark et al. | Aug 2003 | B1 |
6615108 | Peless et al. | Sep 2003 | B1 |
6667592 | Jacobs et al. | Dec 2003 | B2 |
6690134 | Jones et al. | Feb 2004 | B1 |
6764373 | Osawa et al. | Jul 2004 | B1 |
6781338 | Jones et al. | Aug 2004 | B2 |
6809490 | Jones et al. | Oct 2004 | B2 |
6965209 | Jones et al. | Nov 2005 | B2 |
6972834 | Oka et al. | Dec 2005 | B1 |
7155308 | Jones | Dec 2006 | B2 |
7173391 | Jones et al. | Feb 2007 | B2 |
7196487 | Jones et al. | Mar 2007 | B2 |
7302345 | Kwon et al. | Nov 2007 | B2 |
7388343 | Jones et al. | Jun 2008 | B2 |
7389156 | Ziegler et al. | Jun 2008 | B2 |
7448113 | Jones et al. | Nov 2008 | B2 |
7483151 | Zganec et al. | Jan 2009 | B2 |
7507948 | Park et al. | Mar 2009 | B2 |
7539557 | Yamauchi | May 2009 | B2 |
7571511 | Jones et al. | Aug 2009 | B2 |
7636982 | Jones et al. | Dec 2009 | B2 |
7656541 | Waslowski et al. | Feb 2010 | B2 |
7761954 | Ziegler et al. | Jul 2010 | B2 |
7801676 | Kurosawa et al. | Sep 2010 | B2 |
8438695 | Gilbert et al. | May 2013 | B2 |
8594019 | Misumi | Nov 2013 | B2 |
8739355 | Morse et al. | Jun 2014 | B2 |
8843245 | Choe et al. | Sep 2014 | B2 |
8855914 | Alexander et al. | Oct 2014 | B1 |
8892251 | Dooley et al. | Nov 2014 | B1 |
8921752 | Iizuka | Dec 2014 | B2 |
8982217 | Hickman | Mar 2015 | B1 |
9002511 | Hickerson et al. | Apr 2015 | B1 |
9026302 | Stout et al. | May 2015 | B2 |
9037294 | Chung et al. | May 2015 | B2 |
9043017 | Jung et al. | May 2015 | B2 |
9149170 | Ozick et al. | Oct 2015 | B2 |
9220386 | Gilbert, Jr. et al. | Dec 2015 | B2 |
9486924 | Dubrovsky et al. | Nov 2016 | B2 |
9717387 | Szatmary et al. | Aug 2017 | B1 |
10228697 | Yoshino | Mar 2019 | B2 |
20020016649 | Jones | Feb 2002 | A1 |
20020103575 | Sugawara | Aug 2002 | A1 |
20020120364 | Colens | Aug 2002 | A1 |
20030025472 | Jones et al. | Feb 2003 | A1 |
20030030398 | Jacobs et al. | Feb 2003 | A1 |
20030120389 | Abramson et al. | Jun 2003 | A1 |
20030142925 | Melchior et al. | Jul 2003 | A1 |
20040020000 | Jones | Feb 2004 | A1 |
20040049877 | Jones et al. | Mar 2004 | A1 |
20040187457 | Colens | Sep 2004 | A1 |
20040207355 | Jones et al. | Oct 2004 | A1 |
20050000543 | Taylor et al. | Jan 2005 | A1 |
20050010331 | Taylor et al. | Jan 2005 | A1 |
20050041839 | Saitou et al. | Feb 2005 | A1 |
20050067994 | Jones et al. | Mar 2005 | A1 |
20050156562 | Cohen et al. | Jul 2005 | A1 |
20050171636 | Tani | Aug 2005 | A1 |
20050171644 | Tani | Aug 2005 | A1 |
20050204717 | Colens | Sep 2005 | A1 |
20050212680 | Uehigashi | Sep 2005 | A1 |
20050256610 | Orita | Nov 2005 | A1 |
20060012493 | Karlsson et al. | Jan 2006 | A1 |
20060020369 | Taylor | Jan 2006 | A1 |
20060095158 | Lee et al. | May 2006 | A1 |
20060237634 | Kim | Oct 2006 | A1 |
20070027579 | Suzuki et al. | Feb 2007 | A1 |
20070061041 | Zweig | Mar 2007 | A1 |
20070234492 | Svendsen et al. | Oct 2007 | A1 |
20070266508 | Jones et al. | Nov 2007 | A1 |
20070282484 | Chung et al. | Dec 2007 | A1 |
20080046125 | Myeong et al. | Feb 2008 | A1 |
20080140255 | Ziegler et al. | Jun 2008 | A1 |
20080155768 | Ziegler et al. | Jul 2008 | A1 |
20080192256 | Wolf et al. | Aug 2008 | A1 |
20080307590 | Jones et al. | Dec 2008 | A1 |
20090048727 | Hong et al. | Feb 2009 | A1 |
20090051921 | Masahiko | Feb 2009 | A1 |
20090100129 | Vigil | Apr 2009 | A1 |
20090177320 | Lee et al. | Jul 2009 | A1 |
20090182464 | Myeong et al. | Jul 2009 | A1 |
20090281661 | Dooley et al. | Nov 2009 | A1 |
20100030380 | Shah et al. | Feb 2010 | A1 |
20100049365 | Jones et al. | Feb 2010 | A1 |
20100082193 | Chiappetta | Apr 2010 | A1 |
20100257690 | Jones et al. | Oct 2010 | A1 |
20100257691 | Jones et al. | Oct 2010 | A1 |
20100263158 | Jones et al. | Oct 2010 | A1 |
20100324731 | Letsky | Dec 2010 | A1 |
20100324736 | Yoo et al. | Dec 2010 | A1 |
20110054689 | Nielsen et al. | Mar 2011 | A1 |
20110137461 | Kong et al. | Jun 2011 | A1 |
20110194755 | Jeong et al. | Aug 2011 | A1 |
20110211731 | Lee et al. | Sep 2011 | A1 |
20110224824 | Lee et al. | Sep 2011 | A1 |
20110236026 | Yoo et al. | Sep 2011 | A1 |
20110238214 | Yoo et al. | Sep 2011 | A1 |
20110264305 | Choe et al. | Oct 2011 | A1 |
20110278082 | Chung et al. | Nov 2011 | A1 |
20110295420 | Wagner | Dec 2011 | A1 |
20120008128 | Bamji | Jan 2012 | A1 |
20120013907 | Jung et al. | Jan 2012 | A1 |
20120022785 | DiBernardo et al. | Jan 2012 | A1 |
20120060320 | Lee et al. | Mar 2012 | A1 |
20120069457 | Wolf et al. | Mar 2012 | A1 |
20120169497 | Schnittman et al. | Jul 2012 | A1 |
20120173070 | Schnittman | Jul 2012 | A1 |
20120215380 | Fouillade et al. | Aug 2012 | A1 |
20120223216 | Flaherty et al. | Sep 2012 | A1 |
20120265370 | Kim et al. | Oct 2012 | A1 |
20120271502 | Lee et al. | Oct 2012 | A1 |
20120283905 | Nakano et al. | Nov 2012 | A1 |
20130001398 | Wada et al. | Jan 2013 | A1 |
20130024025 | Hsu | Jan 2013 | A1 |
20130166134 | Shitamoto et al. | Jun 2013 | A1 |
20130206177 | Burlutskiy | Aug 2013 | A1 |
20130217421 | Kim | Aug 2013 | A1 |
20130221908 | Tang | Aug 2013 | A1 |
20130261867 | Burnett et al. | Oct 2013 | A1 |
20130265562 | Tang et al. | Oct 2013 | A1 |
20130317944 | Huang et al. | Nov 2013 | A1 |
20140005933 | Fong et al. | Jan 2014 | A1 |
20140098218 | Wu et al. | Apr 2014 | A1 |
20140100693 | Fong et al. | Apr 2014 | A1 |
20140115797 | Duenne | May 2014 | A1 |
20140124004 | Rosenstein et al. | May 2014 | A1 |
20140128093 | Das et al. | May 2014 | A1 |
20140156125 | Song et al. | Jun 2014 | A1 |
20140207280 | Duffley et al. | Jul 2014 | A1 |
20140207281 | Angle et al. | Jul 2014 | A1 |
20140207282 | Angle | Jul 2014 | A1 |
20140218517 | Kim | Aug 2014 | A1 |
20140257563 | Park et al. | Sep 2014 | A1 |
20140257564 | Sun et al. | Sep 2014 | A1 |
20140257565 | Sun et al. | Sep 2014 | A1 |
20140303775 | Oh et al. | Oct 2014 | A1 |
20140316636 | Hong et al. | Oct 2014 | A1 |
20140324270 | Chan et al. | Oct 2014 | A1 |
20140343783 | Lee | Nov 2014 | A1 |
20150115138 | Heng et al. | Apr 2015 | A1 |
20150115876 | Noh et al. | Apr 2015 | A1 |
20150120056 | Noh et al. | Apr 2015 | A1 |
20150151646 | Noiri | Jun 2015 | A1 |
20150168954 | Hickerson et al. | Jun 2015 | A1 |
20150173578 | Kim et al. | Jun 2015 | A1 |
20150202772 | Kim | Jul 2015 | A1 |
20150212520 | Artes | Jul 2015 | A1 |
20150223659 | Han et al. | Aug 2015 | A1 |
20150260829 | Wada | Sep 2015 | A1 |
20150265125 | Lee et al. | Sep 2015 | A1 |
20150269823 | Yamanishi | Sep 2015 | A1 |
20150314453 | Witelson et al. | Nov 2015 | A1 |
20150367513 | Gettings et al. | Dec 2015 | A1 |
20160008982 | Artes et al. | Jan 2016 | A1 |
20160037983 | Hillen et al. | Feb 2016 | A1 |
20160041029 | T'ng et al. | Feb 2016 | A1 |
20160066759 | Langhammer et al. | Mar 2016 | A1 |
20160103451 | Vicenti | Apr 2016 | A1 |
20160123618 | Hester | May 2016 | A1 |
20160132056 | Yoshino | May 2016 | A1 |
20160150933 | Duenne et al. | Jun 2016 | A1 |
20160165795 | Balutis et al. | Jun 2016 | A1 |
20160166126 | Morin et al. | Jun 2016 | A1 |
20160200161 | Van Den Bossche et al. | Jul 2016 | A1 |
20160209217 | Babu et al. | Jul 2016 | A1 |
20160213218 | Ham et al. | Jul 2016 | A1 |
20160229060 | Kim et al. | Aug 2016 | A1 |
20160271795 | Vicenti | Sep 2016 | A1 |
20160278090 | Moeller | Sep 2016 | A1 |
20160282873 | Masaki et al. | Sep 2016 | A1 |
20160297072 | Williams et al. | Oct 2016 | A1 |
20160298970 | Lindhe et al. | Oct 2016 | A1 |
20170000307 | Choi | Jan 2017 | A1 |
20170001311 | Bushman et al. | Jan 2017 | A1 |
20170083022 | Tang | Mar 2017 | A1 |
20170147000 | Hoennige et al. | May 2017 | A1 |
20170177001 | Cao et al. | Jun 2017 | A1 |
20170197314 | Stout et al. | Jul 2017 | A1 |
20170231452 | Saito et al. | Aug 2017 | A1 |
20170364087 | Tang et al. | Dec 2017 | A1 |
Number | Date | Country |
---|---|---|
2015322263 | Apr 2017 | AU |
2322419 | Sep 1999 | CA |
1381340 | Nov 2002 | CN |
1696612 | Nov 2005 | CN |
101945325 | Jan 2011 | CN |
101972129 | Feb 2011 | CN |
102407522 | Apr 2012 | CN |
102738862 | Oct 2012 | CN |
203672362 | Jun 2014 | CN |
104460663 | Mar 2015 | CN |
104634601 | May 2015 | CN |
104765362 | Jul 2015 | CN |
105045098 | Nov 2015 | CN |
105334847 | Feb 2016 | CN |
105467398 | Apr 2016 | CN |
105527619 | Apr 2016 | CN |
105990876 | Oct 2016 | CN |
4421805 | Aug 1995 | DE |
10204223 | Aug 2003 | DE |
10261787 | Jan 2004 | DE |
60002209 | Mar 2004 | DE |
69913150 | Aug 2004 | DE |
102007016802 | May 2008 | DE |
102008028931 | Jun 2008 | DE |
102008014912 | Sep 2009 | DE |
102009059217 | Feb 2011 | DE |
102009041362 | Mar 2011 | DE |
102009052629 | May 2011 | DE |
102010000174 | Jul 2011 | DE |
102010000317 | Aug 2011 | DE |
102010000607 | Sep 2011 | DE |
102010017211 | Dec 2011 | DE |
102010017689 | Jan 2012 | DE |
102010033768 | Feb 2012 | DE |
102011050357 | Feb 2012 | DE |
102012201870 | Aug 2012 | DE |
102011006062 | Sep 2012 | DE |
102011051729 | Jan 2013 | DE |
102012211071 | Nov 2013 | DE |
102012105608 | Jan 2014 | DE |
102012109004 | Mar 2014 | DE |
202014100346 | Mar 2014 | DE |
102012112035 | Jun 2014 | DE |
102012112036 | Jun 2014 | DE |
102013100192 | Jul 2014 | DE |
102014110265 | Jul 2014 | DE |
102014113040 | Sep 2014 | DE |
102013104399 | Oct 2014 | DE |
102013104547 | Nov 2014 | DE |
102015006014 | May 2015 | DE |
102014012811 | Oct 2015 | DE |
102015119501 | Nov 2015 | DE |
102014110104 | Jan 2016 | DE |
102016102644 | Feb 2016 | DE |
102016114594 | Feb 2018 | DE |
102016125319 | Jun 2018 | DE |
0142594 | May 1985 | EP |
0402764 | Dec 1990 | EP |
0769923 | May 1997 | EP |
1062524 | Dec 2000 | EP |
1342984 | Sep 2003 | EP |
1533629 | May 2005 | EP |
1553536 | Jul 2005 | EP |
1557730 | Jul 2005 | EP |
1621948 | Feb 2006 | EP |
1942313 | Jul 2008 | EP |
1947477 | Jul 2008 | EP |
1983396 | Oct 2008 | EP |
2027806 | Feb 2009 | EP |
2053417 | Apr 2009 | EP |
2078996 | Jul 2009 | EP |
2287697 | Feb 2011 | EP |
2327957 | Jun 2011 | EP |
1941411 | Sep 2011 | EP |
2407847 | Jan 2012 | EP |
2450762 | May 2012 | EP |
2457486 | May 2012 | EP |
2498158 | Sep 2012 | EP |
2502539 | Sep 2012 | EP |
2511782 | Oct 2012 | EP |
2515196 | Oct 2012 | EP |
2573639 | Mar 2013 | EP |
2595024 | May 2013 | EP |
2740013 | Jun 2014 | EP |
2741159 | Jun 2014 | EP |
2853976 | Apr 2015 | EP |
2870852 | May 2015 | EP |
3079030 | Nov 2015 | EP |
3156873 | Apr 2017 | EP |
3184013 | Jun 2017 | EP |
2381328 | Jun 2021 | EP |
2509989 | Jul 2014 | GB |
2509990 | Jul 2014 | GB |
2509991 | Jul 2014 | GB |
2513912 | Nov 2014 | GB |
H04338433 | Nov 1992 | JP |
2001125641 | May 2001 | JP |
2002085305 | Mar 2002 | JP |
2003330543 | Nov 2003 | JP |
2004133882 | Apr 2004 | JP |
2005205028 | Aug 2005 | JP |
2009238055 | Oct 2009 | JP |
2010219736 | Sep 2010 | JP |
2010227894 | Oct 2010 | JP |
2011-233149 | Nov 2011 | JP |
2013146302 | Aug 2013 | JP |
2014176260 | Sep 2014 | JP |
201541203 | Mar 2015 | JP |
20070045641 | May 2007 | KR |
100735565 | Jul 2007 | KR |
100815545 | Mar 2008 | KR |
20110092158 | Aug 2011 | KR |
20140073854 | Jun 2014 | KR |
20140145648 | Dec 2014 | KR |
20150009413 | Jan 2015 | KR |
20150050161 | May 2015 | KR |
20150086075 | Jul 2015 | KR |
20170004343 | Jul 2015 | KR |
20150124011 | Nov 2015 | KR |
20150124013 | Nov 2015 | KR |
20150124014 | Nov 2015 | KR |
20150127937 | Nov 2015 | KR |
101640706 | Jul 2016 | KR |
20160097051 | Aug 2016 | KR |
9523346 | Aug 1995 | WO |
9928800 | Jun 1999 | WO |
200004430 | Jan 2000 | WO |
2005074362 | Aug 2005 | WO |
2007028667 | Mar 2007 | WO |
2012099694 | Jul 2012 | WO |
2012157951 | Nov 2012 | WO |
2014017256 | Jan 2014 | WO |
2014043732 | Mar 2014 | WO |
2014055966 | Apr 2014 | WO |
2014113091 | Jul 2014 | WO |
2015018437 | Feb 2015 | WO |
2015025599 | Feb 2015 | WO |
2015072897 | May 2015 | WO |
2015082017 | Jun 2015 | WO |
2015090398 | Jun 2015 | WO |
2015158240 | Oct 2015 | WO |
2015181995 | Dec 2015 | WO |
2016019996 | Feb 2016 | WO |
2016027957 | Feb 2016 | WO |
2016028021 | Feb 2016 | WO |
2016031702 | Mar 2016 | WO |
2016048077 | Mar 2016 | WO |
2016050215 | Apr 2016 | WO |
2016091312 | Jun 2016 | WO |
2016095966 | Jun 2016 | WO |
Entry |
---|
English Translation for reference KR20070045641A, (Year: 2007). |
English translation KR20170004343A (Year: 2015). |
English translation JP2010219736 (Year: 2010). |
Konolige et al., “A Low-Cost Laser Distance Sensor,” 2008 IEEE International Conference on Robotics and Automation, Pasadena, CA, USA, May 19-23, 2008, document of 7 pages. |
Oh et al., “Autonomous Battery Recharging for Indoor Mobile Robots,” Massachusetts Institute of Technology Press, Aug. 30, 2000, document of 6 pages, XP055321836. |
Siegwart, “Introduction to Autonomous Mobile Robots”, Massachusetts, ISBN 978-0-26-219502-7, (2004), pp. 104-115, 151-163, 250-251, document of 37 pages. http://www.robotee.com/EBooks/Introduction_to_Autonomous_Mobile_Robots.pdf, XP055054850. |
Lymberopoulos et al., “A Realistic Evaluation and Comparison of Indoor Location Technologies: Experiences and Lessons Learned,” IPSN '15, Apr. 14-16, 2015, Seattle, WA, USA, document of 12 pages. http://dx.doi.org/10.1145/2737095.27. |
Patent Cooperation Treaty, “International Search Report” and English translation thereof, issued in International Application No. PCT/EP2016/080235, by Examiner Patrick Kielhöfer, document of 8 pages, dated Jun. 6, 2017. |
Thrun et al., “Probabilistic Robotics”; 1999, 492 pages. |
Choset et al., “Principles of Robot Motion”, Theory, Algorithms, and Implementations, Chapter 6—Cell Decompositions, 2004, document of 41 pages. |
Durrant-Whyte et al., “Simultaneous Localization and Mapping (SLAM): Part I The Essential Algorithms”, in: IEEE Robotics and Automation Magazine, vol. 13, No. 2, pp. 99-108 Jun. 2006. |
Vasquez-Gomez et al., “View planning for 3D object reconstruction with a mobile manipulator robot,” 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems, Sep. 14, 2014 IEEE, pp. 4227-4233. |
Forlizzi, How robotic products become social products: An ethnographic study of cleaning in the home, 2007, IEEE, p. 129-136 (Year: 2007). |
Kim et al., “User-Centered Approach to Path Planning of Cleaning Robots: Analyzing Users Cleaning Behavior.” Proceedings of the 2007 ACM/IEEE Conference on Human-Robot Interaction, Mar. 8-11, 2007, pp. 373-380. |
Mahyuddin et al., “Neuro-fuzzy algorithm implemented in Altera's FPGA for mobile robot's obstacle avoidance mission”, Tencon 2009—2009 IEEE Region 10 Conference, IEEE, Piscataway, NJ, USA, Jan. 23, 2009; document of 6 pages. |
Neto et al., Human-Machine Interface Based on Electro-Biological Signals for Mobile Vehicles, 2006, IEEE, p. 2954-2959 (Year: 2006). |
Sick Sensor Intelligence, “LMS200/211/221/291 Laser Measurement Systems”, Jan. 2007, pp. 1-48, XP055581229, http://sicktoolbox.sourceforge.net/docs/sick-lms-technical-description.pdf. |
Japanese Patent Office, “Notice of Refusal,” issued in Japanese patent application No. 2018-5300086, dated Nov. 25, 2021, document of 8 pages. |
Number | Date | Country | |
---|---|---|---|
20180292827 A1 | Oct 2018 | US |
Number | Date | Country | |
---|---|---|---|
Parent | PCT/EP2016/080235 | Dec 2016 | US |
Child | 16005013 | US |