Assignment of a motorized personal assistance apparatus

Information

  • Patent Grant
  • 10287149
  • Patent Number
    10,287,149
  • Date Filed
    Friday, March 4, 2016
    8 years ago
  • Date Issued
    Tuesday, May 14, 2019
    4 years ago
Abstract
An apparatus comprises a plurality of motorized mobile retail facility personal assistance apparatuses and a central computer system. The central computer system includes a network interface that permits the central computer system to wirelessly communicate with the mobile apparatuses and also with at least one user interface unit. Upon receiving a request to use one of the mobile apparatuses, the central computer system assigns an available apparatuses for use by a user during a usage session. This activity can include interfacing with the user to determine one or more specifics regarding their current needs in these regards. The specifics of their needs, in turn, can provide a basis for assigning a particular apparatus from amongst a plurality of differently purposed and differently configured apparatuses.
Description
TECHNICAL FIELD

These teachings relate generally to shopping environments and more particularly to devices, systems and methods for assisting customers and/or workers in those shopping environments.


BACKGROUND

In a modern retail store environment, there is a need to improve the customer experience and/or convenience for the customer. Whether shopping in a large format (big box) store or smaller format (neighborhood) store, customers often require assistance that employees of the store are not always able to provide. For example, particularly during peak hours, there may not be enough employees available to assist customers such that customer questions go unanswered. Additionally, due to high employee turnover rates, available employees may not be fully trained or have access to information to adequately support customers. Other routine tasks also are difficult to keep up with, particularly during peak hours. For example, shopping carts are left abandoned, aisles become messy, inventory is not displayed in the proper locations or is not even placed on the sales floor, shelf prices may not be properly set, and theft is hard to discourage. All of these issues can result in low customer satisfaction or reduced convenience to the customer. With increasing competition from non-traditional shopping mechanisms, such as online shopping provided by e-commerce merchants and alternative store formats, it can be important for “brick and mortar” retailers to focus on improving the overall customer experience and/or convenience.


While offering great convenience in the form of one-stop shopping, such stores can also be at least mildly confusing to some consumers due in large part to the very size of the facility and the plethora of items being offered. That wide variety of items can also present a corresponding issue for some consumers in the form of a shopping cart that becomes so filled with various items that pushing and navigating the cart becomes potentially challenging.


Motorized shopping carts are known and typically comprise a shopping cart having an electric motor to power one or more wheels and navigational controls to permit the user to steer the cart. Motorized shopping carts typically include a seat and have a rechargeable battery that can be charged during periods of non-usage. These devices are typically provided to serve persons with permanent or temporary physical disabilities who may have difficulty walking through a large store or pushing a regular cart.


Though certainly useful, existing approaches in these regards do not address all needs of all potential users.





BRIEF DESCRIPTION OF THE DRAWINGS

The above needs are at least partially met through provision of the approaches described in the following detailed description as regards assigning a motorized personal assistance apparatus, particularly when studied in conjunction with the drawings, wherein:



FIG. 1 comprises a block diagram of a shopping assistance system as configured in accordance with various embodiments of these teachings;



FIGS. 2A and 2B are illustrations of a motorized transport unit of the system of FIG. 1 in a retracted orientation and an extended orientation in accordance with some embodiments;



FIGS. 3A and 3B are illustrations of the motorized transport unit of FIGS. 2A and 2B detachably coupling to a movable item container, such as a shopping cart, in accordance with some embodiments;



FIG. 4 comprises a block diagram of a motorized transport unit as configured in accordance with various embodiments of these teachings;



FIG. 5 comprises a block diagram of a computer device as configured in accordance with various embodiments of these teachings;



FIG. 6 comprises a flow diagram as configured in accordance with various embodiments of these teachings;



FIG. 7 comprises a flow diagram as configured in accordance with various embodiments of these teachings;



FIG. 8 comprises a screen shot as configured in accordance with various embodiments of these teachings;



FIG. 9 comprises a flow diagram as configured in accordance with various embodiments of these teachings;



FIG. 10 comprises a blow diagram as configured in accordance with various embodiments of these teachings; and



FIG. 11 comprises a flow diagram as configured in accordance with various embodiments of these teachings.





Elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions and/or relative positioning of some of the elements in the figures may be exaggerated relative to other elements to help to improve understanding of various embodiments of the present teachings. Also, common but well-understood elements that are useful or necessary in a commercially feasible embodiment are often not depicted in order to facilitate a less obstructed view of these various embodiments of the present teachings. Certain actions and/or steps may be described or depicted in a particular order of occurrence while those skilled in the art will understand that such specificity with respect to sequence is not actually required. The terms and expressions used herein have the ordinary technical meaning as is accorded to such terms and expressions by persons skilled in the technical field as set forth above except where different specific meanings have otherwise been set forth herein.


DETAILED DESCRIPTION

The following description is not to be taken in a limiting sense, but is made merely for the purpose of describing the general principles of exemplary embodiments. Reference throughout this specification to “one embodiment,” “an embodiment,” or similar language means that a particular feature, structure, or characteristic described in connection with the embodiment is included in at least one embodiment of the present invention. Thus, appearances of the phrases “in one embodiment,” “in an embodiment,” and similar language throughout this specification may, but do not necessarily, all refer to the same embodiment.


Generally speaking, pursuant to various embodiments, systems, devices and methods are provided for assistance of persons at a shopping facility. Generally, assistance may be provided to customers or shoppers at the facility and/or to workers at the facility. The facility may be any type of shopping facility at a location in which products for display and/or for sale are variously distributed throughout the shopping facility space. The shopping facility may be a retail sales facility, or any other type of facility in which products are displayed and/or sold. The shopping facility may include one or more of sales floor areas, checkout locations, parking locations, entrance and exit areas, stock room areas, stock receiving areas, hallway areas, common areas shared by merchants, and so on. Generally, a shopping facility includes areas that may be dynamic in terms of the physical structures occupying the space or area and objects, items, machinery and/or persons moving in the area. For example, the shopping area may include product storage units, shelves, racks, modules, bins, etc., and other walls, dividers, partitions, etc. that may be configured in different layouts or physical arrangements. In other example, persons or other movable objects may be freely and independently traveling through the shopping facility space. And in other example, the persons or movable objects move according to known travel patterns and timing. The facility may be any size of format facility, and may include products from one or more merchants. For example, a facility may be a single store operated by one merchant or may be a collection of stores covering multiple merchants such as a mall.


Generally, the system makes use of automated, robotic mobile devices, e.g., motorized transport units, that are capable of self-powered movement through a space of the shopping facility and providing any number of functions. Movement and operation of such devices may be controlled by a central computer system or may be autonomously controlled by the motorized transport units themselves. Various embodiments provide one or more user interfaces to allow various users to interact with the system including the automated mobile devices and/or to directly interact with the automated mobile devices. In some embodiments, the automated mobile devices and the corresponding system serve to enhance a customer shopping experience in the shopping facility, e.g., by assisting shoppers and/or workers at the facility.


Generally speaking, pursuant to these various embodiments, an apparatus comprises a plurality of motorized mobile retail facility personal assistance apparatuses and a central computer system. By one approach the central computer system includes a network interface that permits the central computer system to wirelessly communicate with the plurality of motorized mobile retail facility personal assistance apparatuses and also with at least one user interface unit.


The central computer system is configured to receive a request from a user (via the at least one user interface unit referred to above) to use one of the plurality of motorized mobile retail facility personal assistance apparatuses. In response to receiving such a request, the central computer system assigns an available one of the plurality of motorized mobile retail facility personal assistance apparatuses for use by that user during a usage session.


By one approach that assignment includes, at least in part, wirelessly transmitting to the user via the user interface unit a globally unique session token that uniquely corresponds to the usage session. If desired, the user interface unit then employs that globally unique session token as an encryption key for at least some wireless communications during at least part of the aforementioned usage session. This can include communications with the assigned motorized mobile retail facility personal assistance apparatus and/or with the central computer system as desired.


By another approach, in lieu of the foregoing or in combination therewith, the central computer system is further configured to use that globally unique session token to log tracking information of the user's use of the motorized mobile retail facility personal assistance apparatus during the usage session. That tracking information can include, for example, information regarding activity of the motorized mobile retail facility personal assistance apparatus and/or activity of the user.


These teachings are highly flexible in practice and will accommodate a variety of modifications or approaches as regards the foregoing. For example, the motorized mobile retail facility personal assistance apparatuses may integrally include an item container or may, if desired, comprise a motorized transport device that automatically connects, when assigned by the central computer system, to a movable item container such that the motorized transport device will move the motorized item container to a location of convenience as corresponds to the user during the usage session such that the user can readily place a selected item into the movable item container to facilitate purchasing the selected item.


By one approach the aforementioned user interface unit can comprise, for example, a smart phone or a pad/tablet-styled computer as are known in the art. These teachings will accommodate the user interface unit being mounted to a movable item container (such as a shopping cart) or the user interface unit comprising a user-possessed personally-portable two-way communications unit.


As another example of the flexibility of these teachings, by one approach, when assigning an available motorized mobile retail facility personal assistant apparatuses, the central computer system can also wirelessly transmit to the user interface unit a temporary N-digit code (where “N” is a non-zero integer). In such a case, the user interface unit can be configured to display that code and to provide the user with an opportunity to themselves enter that code into the user interface unit. The central computer system can be configured to determine whether the user interface unit returns that correctly-entered code within a pre-determined period of time. When the code is returned within that pre-determined period of time, the central computer system can then begin the usage session as corresponds to the assignment of the particular motorized mobile retail facility personal assistance apparatus to this user. When, however, the pre-determined period of time expires with the user having failed to enter the temporary N-digit code into the user interface unit, the central computer system can withdraw the assignment of the motorized mobile retail facility personal assistance apparatus for use by this particular user.


If desired, before assigning a motorized mobile retail facility personal assistance apparatus to a particular user, the central computer system can be configured to assess the user's request. This assessment can include, for example, determining whether the user is personally barred from being assigned a motorized mobile retail facility personal assistance apparatus (such barring being based, for example, upon prior misuse of such an apparatus). As another example, this assessment can be based, at least in part, upon a present physical location of the user.


It is possible that an assigned motorized mobile retail facility personal assistance apparatus will be unable to complete a usage session for a particular user. As one simple example, the device's rechargeable power source may be too depleted to complete the assignment. Accordingly, if desired, the central computer system can be configured to arrange to transfer, during a usage session, the assignment of a replacement motorized mobile retail facility personal assistance apparatus for use by the user. Such a transfer can include, for example, transferring the aforementioned globally unique session token to the replacement motorized mobile retail facility personal assistance apparatus.


It is also possible that a user will attempt to enter an area of the store that is off limits to unauthorized persons (such as a back stock room). If desired, the central computer system can be configured to terminate a usage session when the user enters an area that is off-limits to thereby at least prevent the assigned motorized mobile retail facility personal assistance apparatus from entering the off-limits area.


By one approach, in lieu of much of the foregoing or in combination therewith, these teachings will accommodate interfacing with a user to ascertain specific needs of the user to thereby facilitate tasking a particular one of a plurality of motorized transport units from amongst a plurality of differently purposed and differently physically configured motorized transport units to assist the user during their visit to the retail shopping facility. Different motorized transport units may be differently purposed/configured, for example, to serve as disability mobility platforms, grocery carts, pallets, and child-carrying platforms, to note but a few examples in these regards.


These teachings provide an effective and efficient way to ensure an appropriate assignment of a motorized mobile retail facility personal assistance apparatus to a given user. By one approach the assignment process is fully automated with the exception of specific actions that are appropriately taken by the user. Providing an effective assignment mechanism, in turn, can help to avoid confusion, misuse, mistakes, and other undesired results.


These and other benefits may become clearer upon making a thorough review and study of the following detailed description. Referring now to the drawings, FIG. 1 illustrates embodiments of a shopping facility assistance system 100 that can serve to carry out at least some of the teachings set forth herein. It will be understood that the details of this example are intended to serve in an illustrative capacity and are not necessarily intended to suggest any limitations as regards the present teachings. It is noted that generally, FIGS. 1-5 describe the general functionality of several embodiments of a system, and FIGS. 6-8 expand on some functionalities of some embodiments of the system and/or embodiments independent of such systems.


In the example of FIG. 1, a shopping assistance system 100 is implemented in whole or in part at a shopping facility 101. Generally, the system 100 includes one or more motorized transport units (MTUs) 102; one or more item containers 104; a central computer system 106 having at least one control circuit 108, at least one memory 110 and at least one network interface 112; at least one user interface unit 114; a location determination system 116; at least one video camera 118; at least one motorized transport unit (MTU) dispenser 120; at least one motorized transport unit (MTU) docking station 122; at least one wireless network 124; at least one database 126; at least one user interface computer device 128; an item display module 130; and a locker or an item storage unit 132. It is understood that more or fewer of such components may be included in different embodiments of the system 100.


These motorized transport units 102 are located in the shopping facility 101 and are configured to move throughout the shopping facility space. Further details regarding such motorized transport units 102 appear further below. Generally speaking, these motorized transport units 102 are configured to either comprise, or to selectively couple to, a corresponding movable item container 104. A simple example of an item container 104 would be a shopping cart as one typically finds at many retail facilities, or a rocket cart, a flatbed cart or any other mobile basket or platform that may be used to gather items for potential purchase.


In some embodiments, these motorized transport units 102 wirelessly communicate with, and are wholly or largely controlled by, the central computer system 106. In particular, in some embodiments, the central computer system 106 is configured to control movement of the motorized transport units 102 through the shopping facility space based on a variety of inputs. For example, the central computer system 106 communicates with each motorized transport unit 102 via the wireless network 124 which may be one or more wireless networks of one or more wireless network types (such as, a wireless local area network, a wireless personal area network, a wireless mesh network, a wireless star network, a wireless wide area network, a cellular network, and so on), capable of providing wireless coverage of the desired range of the motorized transport units 102 according to any known wireless protocols, including but not limited to a cellular, Wi-Fi, Zigbee or Bluetooth network.


By one approach the central computer system 106 is a computer based device and includes at least one control circuit 108, at least one memory 110 and at least one wired and/or wireless network interface 112. Such a control circuit 108 can comprise a fixed-purpose hard-wired platform or can comprise a partially or wholly programmable platform, such as a microcontroller, an application specification integrated circuit, a field programmable gate array, and so on. These architectural options are well known and understood in the art and require no further description here. This control circuit 108 is configured (for example, by using corresponding programming stored in the memory 110 as will be well understood by those skilled in the art) to carry out one or more of the steps, actions, and/or functions described herein.


In this illustrative example the control circuit 108 operably couples to one or more memories 110. The memory 110 may be integral to the control circuit 108 or can be physically discrete (in whole or in part) from the control circuit 108 as desired. This memory 110 can also be local with respect to the control circuit 108 (where, for example, both share a common circuit board, chassis, power supply, and/or housing) or can be partially or wholly remote with respect to the control circuit 108 (where, for example, the memory 110 is physically located in another facility, metropolitan area, or even country as compared to the control circuit 108).


This memory 110 can serve, for example, to non-transitorily store the computer instructions that, when executed by the control circuit 108, cause the control circuit 108 to behave as described herein. (As used herein, this reference to “non-transitorily” will be understood to refer to a non-ephemeral state for the stored contents (and hence excludes when the stored contents merely constitute signals or waves) rather than volatility of the storage media itself and hence includes both non-volatile memory (such as read-only memory (ROM) as well as volatile memory (such as an erasable programmable read-only memory (EPROM).)


Additionally, at least one database 126 may be accessible by the central computer system 106. Such databases may be integrated into the central computer system 106 or separate from it. Such databases may be at the location of the shopping facility 101 or remote from the shopping facility 101. Regardless of location, the databases comprise memory to store and organize certain data for use by the central control system 106. In some embodiments, the at least one database 126 may store data pertaining to one or more of: shopping facility mapping data, customer data, customer shopping data and patterns, inventory data, product pricing data, and so on.


In this illustrative example, the central computer system 106 also wirelessly communicates with a plurality of user interface units 114. These teachings will accommodate a variety of user interface units including, but not limited to, mobile and/or handheld electronic devices such as so-called smart phones and portable computers such as tablet/pad-styled computers. Generally speaking, these user interface units 114 should be able to wirelessly communicate with the central computer system 106 via a wireless network, such as the wireless network 124 of the shopping facility 101 (such as a Wi-Fi wireless network). These user interface units 114 generally provide a user interface for interaction with the system. In some embodiments, a given motorized transport unit 102 is paired with, associated with, assigned to or otherwise made to correspond with a given user interface unit 114. In some embodiments, these user interface units 114 should also be able to receive verbally-expressed input from a user and forward that content to the central computer system 106 or a motorized transport unit 102 and/or convert that verbally-expressed input into a form useful to the central computer system 106 or a motorized transport unit 102.


By one approach at least some of the user interface units 114 belong to corresponding customers who have come to the shopping facility 101 to shop. By another approach, in lieu of the foregoing or in combination therewith, at least some of the user interface units 114 belong to the shopping facility 101 and are loaned to individual customers to employ as described herein. In some embodiments, one or more user interface units 114 are attachable to a given movable item container 104 or are integrated with the movable item container 104. Similarly, in some embodiments, one or more user interface units 114 may be those of shopping facility workers, belong to the shopping facility 101 and are loaned to the workers, or a combination thereof.


In some embodiments, the user interface units 114 may be general purpose computer devices that include computer programming code to allow it to interact with the system 106. For example, such programming may be in the form of an application installed on the user interface unit 114 or in the form of a browser that displays a user interface provided by the central computer system 106 or other remote computer or server (such as a web server). In some embodiments, one or more user interface units 114 may be special purpose devices that are programmed to primarily function as a user interface for the system 100. Depending on the functionality and use case, user interface units 114 may be operated by customers of the shopping facility or may be operated by workers at the shopping facility, such as facility employees (associates or colleagues), vendors, suppliers, contractors, etc.


By one approach, the system 100 optionally includes one or more video cameras 118. Captured video imagery from such a video camera 118 can be provided to the central computer system 106. That information can then serve, for example, to help the central computer system 106 determine a present location of one or more of the motorized transport units 102 and/or determine issues or concerns regarding automated movement of those motorized transport units 102 in the shopping facility space. As one simple example in these regards, such video information can permit the central computer system 106, at least in part, to detect an object in a path of movement of a particular one of the motorized transport units 102.


By one approach these video cameras 118 comprise existing surveillance equipment employed at the shopping facility 101 to serve, for example, various security purposes. By another approach these video cameras 118 are dedicated to providing video content to the central computer system 106 to facilitate the latter's control of the motorized transport units 102. If desired, the video cameras 118 can have a selectively movable field of view and/or zoom capability that the central computer system 106 controls as appropriate to help ensure receipt of useful information at any given moment.


In some embodiments, a location detection system 116 is provided at the shopping facility 101. The location detection system 116 provides input to the central computer system 106 useful to help determine the location of one or more of the motorized transport units 102. In some embodiments, the location detection system 116 includes a series of light sources (e.g., LEDs (light-emitting diodes)) that are mounted in the ceiling at known positions throughout the space and that each encode data in the emitted light that identifies the source of the light (and thus, the location of the light). As a given motorized transport unit 102 moves through the space, light sensors (or light receivers) at the motorized transport unit 102, on the movable item container 104 and/or at the user interface unit 114 receive the light and can decode the data. This data is sent back to the central computer system 106 which can determine the position of the motorized transport unit 102 by the data of the light it receives, since it can relate the light data to a mapping of the light sources to locations at the facility 101. Generally, such lighting systems are known and commercially available, e.g., the ByteLight system from ByteLight of Boston, Mass. In embodiments using a ByteLight system, a typical display screen of the typical smart phone device can be used as a light sensor or light receiver to receive and process data encoded into the light from the ByteLight light sources.


In other embodiments, the location detection system 116 includes a series of low energy radio beacons (e.g., Bluetooth low energy beacons) at known positions throughout the space and that each encode data in the emitted radio signal that identifies the beacon (and thus, the location of the beacon). As a given motorized transport unit 102 moves through the space, low energy receivers at the motorized transport unit 102, on the movable item container 104 and/or at the user interface unit 114 receive the radio signal and can decode the data. This data is sent back to the central computer system 106 which can determine the position of the motorized transport unit 102 by the location encoded in the radio signal it receives, since it can relate the location data to a mapping of the low energy radio beacons to locations at the facility 101. Generally, such low energy radio systems are known and commercially available. In embodiments using a Bluetooth low energy radio system, a typical Bluetooth radio of a typical smart phone device can be used as a receiver to receive and process data encoded into the Bluetooth low energy radio signals from the Bluetooth low energy beacons.


In still other embodiments, the location detection system 116 includes a series of audio beacons at known positions throughout the space and that each encode data in the emitted audio signal that identifies the beacon (and thus, the location of the beacon). As a given motorized transport unit 102 moves through the space, microphones at the motorized transport unit 102, on the movable item container 104 and/or at the user interface unit 114 receive the audio signal and can decode the data. This data is sent back to the central computer system 106 which can determine the position of the motorized transport unit 102 by the location encoded in the audio signal it receives, since it can relate the location data to a mapping of the audio beacons to locations at the facility 101. Generally, such audio beacon systems are known and commercially available. In embodiments using an audio beacon system, a typical microphone of a typical smart phone device can be used as a receiver to receive and process data encoded into the audio signals from the audio beacon.


Also optionally, the central computer system 106 can operably couple to one or more user interface computers 128 (comprising, for example, a display and a user input interface such as a keyboard, touch screen, and/or cursor-movement device). Such a user interface computer 128 can permit, for example, a worker (e.g., an associate, analyst, governmental inspector, etc.) at the retail or shopping facility 101 to monitor the operations of the central computer system 106 and/or to attend to any of a variety of administrative, configuration or evaluation tasks as may correspond to the programming and operation of the central computer system 106. Such user interface computers 128 may be at or remote from the location of the facility 101 and may access one or more the databases 126.


In some embodiments, the system 100 includes at least one motorized transport unit (MTU) storage unit or dispenser 120 at various locations in the shopping facility 101. The dispenser 120 provides for storage of motorized transport units 102 that are ready to be assigned to customers and/or workers. In some embodiments, the dispenser 120 takes the form of a cylinder within which motorized transports units 102 are stacked and released through the bottom of the dispenser 120. Further details of such embodiments are provided further below. In some embodiments, the dispenser 120 may be fixed in location or may be mobile and capable of transporting itself to a given location or utilizing a motorized transport unit 102 to transport the dispenser 120, then dispense one or more motorized transport units 102.


In some embodiments, the system 100 includes at least one motorized transport unit (MTU) docking station 122. These docking stations 122 provide locations where motorized transport units 102 can travel and connect to. For example, the motorized transport units 102 may be stored and charged at the docking station 122 for later use, and/or may be serviced at the docking station 122.


In accordance with some embodiments, a given motorized transport unit 102 detachably connects to a movable item container 104 and is configured to move the movable item container 104 through the shopping facility space under control of the central computer system 106 and/or the user interface unit 114. For example, a motorized transport unit 102 can move to a position underneath a movable item container 104 (such as a shopping cart, a rocket cart, a flatbed cart, or any other mobile basket or platform), align itself with the movable item container 104 (e.g., using sensors) and then raise itself to engage an undersurface of the movable item container 104 and lift a portion of the movable item container 104. Once the motorized transport unit is cooperating with the movable item container 104 (e.g., lifting a portion of the movable item container), the motorized transport unit 102 can continue to move throughout the facility space 101 taking the movable item container 104 with it.


In some examples, the motorized transport unit 102 takes the form of the motorized transport unit 202 of FIGS. 2A-3B as it engages and detachably connects to a given movable item container 104. It is understood that in other embodiments, the motorized transport unit 102 may not lift a portion of the movable item container 104, but that it removably latches to, connects to or otherwise attaches to a portion of the movable item container 104 such that the movable item container 104 can be moved by the motorized transport unit 102. For example, the motorized transport unit 102 can connect to a given movable item container using a hook, a mating connector, a magnet, and so on.


In addition to detachably coupling to movable item containers 104 (such as shopping carts), in some embodiments, motorized transport units 102 can move to and engage or connect to an item display module 130 and/or an item storage unit or locker 132. For example, an item display module 130 may take the form of a mobile display rack or shelving unit configured to house and display certain items for sale. It may be desired to position the display module 130 at various locations within the shopping facility 101 at various times. Thus, one or more motorized transport units 102 may move (as controlled by the central computer system 106) underneath the item display module 130, extend upward to lift the module 130 and then move it to the desired location. A storage locker 132 may be a storage device where items for purchase are collected and placed therein for a customer and/or worker to later retrieve. In some embodiments, one or more motorized transport units 102 may be used to move the storage locker to a desired location in the shopping facility 101. Similar to how a motorized transport unit engages a movable item container 104 or item display module 130, one or more motorized transport units 102 may move (as controlled by the central computer system 106) underneath the storage locker 132, extend upward to lift the locker 132 and then move it to the desired location.



FIGS. 2A and 2B illustrate some embodiments of a motorized transport unit 202, similar to the motorized transport unit 102 shown in the system of FIG. 1. In this embodiment, the motorized transport unit 202 takes the form of a disc-shaped robotic device having motorized wheels (not shown), a lower body portion 204 and an upper body portion 206 that fits over at least part of the lower body portion 204. It is noted that in other embodiments, the motorized transport unit may have other shapes and/or configurations, and is not limited to disc-shaped. For example, the motorized transport unit may be cubic, octagonal, triangular, or other shapes, and may be dependent on a movable item container with which the motorized transport unit is intended to cooperate. Also included are guide members 208.


In FIG. 2A, the motorized transport unit 202 is shown in a retracted position in which the upper body portion 206 fits over the lower body portion 204 such that the motorized transport unit 202 is in its lowest profile orientation which is generally the preferred orientation for movement when it is unattached to a movable item container 104 for example.


In FIG. 2B, the motorized transport unit 202 is shown in an extended position in which the upper body portion 206 is moved upward relative to the lower body portion 204 such that the motorized transport unit 202 is in its highest profile orientation for movement when it is lifting and attaching to a movable item container 104 for example. The mechanism within the motorized transport unit 202 is designed to provide sufficient lifting force to lift the weight of the upper body portion 206 and other objects to be lifted by the motorized transport unit 202, such as movable item containers 104 and items placed within the movable item container, item display modules 130 and items supported by the item display module, and storage lockers 132 and items placed within the storage locker.


The guide members 208 are embodied as pegs or shafts that extend horizontally from the both the upper body portion 206 and the lower body portion 204. In some embodiments, these guide members 208 assist docking the motorized transport unit 202 to a docking station 122 or a dispenser 120.


In some embodiments, the lower body portion 204 and the upper body portion are capable to moving independently of each other. For example, the upper body portion 206 may be raised and/or rotated relative to the lower body portion 204. That is, one or both of the upper body portion 206 and the lower body portion 204 may move toward/away from the other or rotated relative to the other.


In some embodiments, in order to raise the upper body portion 206 relative to the lower body portion 204, the motorized transport unit 202 includes an internal lifting system (e.g., including one or more electric actuators or rotary drives or motors). Numerous examples of such motorized lifting and rotating systems are known in the art. Accordingly, further elaboration in these regards is not provided here for the sake of brevity.



FIGS. 3A and 3B illustrate some embodiments of the motorized transport unit 202 detachably engaging a movable item container embodied as a shopping cart 302. In FIG. 3A, the motorized transport unit 202 is in the orientation of FIG. 2A such that it is retracted and able to move in position underneath a portion of the shopping cart 302. Once the motorized transport unit 202 is in position (e.g., using sensors), as illustrated in FIG. 3B, the motorized transport unit 202 is moved to the extended position of FIG. 2B such that the front portion 304 of the shopping cart is lifted off of the ground by the motorized transport unit 202, with the wheels 306 at the rear of the shopping cart 302 remaining on the ground. In this orientation, the motorized transport unit 202 is able to move the shopping cart 302 throughout the shopping facility. It is noted that in these embodiments, the motorized transport unit 202 does not bear the weight of the entire cart 302 since the rear wheels 306 rest on the floor. It is understood that in some embodiments, the motorized transport unit 202 may be configured to detachably engage other types of movable item containers, such as rocket carts, flatbed carts or other mobile baskets or platforms.



FIG. 4 presents a more detailed example of some embodiments of the motorized transport unit 102 of FIG. 1. In this example, the motorized transport unit 102 has a housing 402 that contains (partially or fully) or at least supports and carries a number of components. These components include a control unit 404 comprising a control circuit 406 that, like the control circuit 108 of the central computer system 106, controls the general operations of the motorized transport unit 102. Accordingly, the control unit 404 also includes a memory 408 coupled to the control circuit 406 and that stores, for example, operating instructions and/or useful data.


The control circuit 406 operably couples to a motorized wheel system 410. This motorized wheel system 410 functions as a locomotion system to permit the motorized transport unit 102 to move within the aforementioned retail or shopping facility 101 (thus, the motorized wheel system 410 may more generically be referred to as a locomotion system). Generally speaking, this motorized wheel system 410 will include at least one drive wheel (i.e., a wheel that rotates (around a horizontal axis) under power to thereby cause the motorized transport unit 102 to move through interaction with, for example, the floor of the shopping facility 101). The motorized wheel system 410 can include any number of rotating wheels and/or other floor-contacting mechanisms as may be desired and/or appropriate to the application setting.


The motorized wheel system 410 also includes a steering mechanism of choice. One simple example in these regards comprises one or more of the aforementioned wheels that can swivel about a vertical axis to thereby cause the moving motorized transport unit 102 to turn as well.


Numerous examples of motorized wheel systems are known in the art. Accordingly, further elaboration in these regards is not provided here for the sake of brevity save to note that the aforementioned control circuit 406 is configured to control the various operating states of the motorized wheel system 410 to thereby control when and how the motorized wheel system 410 operates.


In this illustrative example, the control circuit 406 also operably couples to at least one wireless transceiver 412 that operates according to any known wireless protocol. This wireless transceiver 412 can comprise, for example, a Wi-Fi-compatible and/or Bluetooth-compatible transceiver that can communicate with the aforementioned central computer system 106 via the aforementioned wireless network 124 of the shopping facility 101. So configured the control circuit 406 of the motorized transport unit 102 can provide information to the central computer system 106 and can receive information and/or instructions from the central computer system 106. As one simple example in these regards, the control circuit 406 can receive instructions from the central computer system 106 regarding movement of the motorized transport unit 102.


These teachings will accommodate using any of a wide variety of wireless technologies as desired and/or as may be appropriate in a given application setting. These teachings will also accommodate employing two or more different wireless transceivers 412 if desired.


The control circuit 406 also couples to one or more on-board sensors 414. These teachings will accommodate a wide variety of sensor technologies and form factors. By one approach at least one such sensor 414 can comprise a light sensor or light receiver. When the aforementioned location detection system 116 comprises a plurality of light emitters disposed at particular locations within the shopping facility 101, such a light sensor can provide information that the control circuit 406 and/or the central computer system 106 employs to determine a present location and/or orientation of the motorized transport unit 102.


As another example, such a sensor 414 can comprise a distance measurement unit configured to detect a distance between the motorized transport unit 102 and one or more objects or surfaces around the motorized transport unit 102 (such as an object that lies in a projected path of movement for the motorized transport unit 102 through the shopping facility 101). These teachings will accommodate any of a variety of distance measurement units including optical units and sound/ultrasound units.


In one example, a sensor 414 comprises a laser distance sensor device capable of determining a distance to objects in proximity to the sensor. In some embodiments, a sensor 414 comprises an optical based scanning device to sense and read optical patterns in proximity to the sensor, such as bar codes variously located on structures in the shopping facility 101. In some embodiments, a sensor 414 comprises a radio frequency identification (RFID) tag reader capable of reading RFID tags in proximity to the sensor. Such sensors may be useful to determine proximity to nearby objects, avoid collisions, orient the motorized transport unit at a proper alignment orientation to engage a movable item container, and so on.


The foregoing examples are intended to be illustrative and are not intended to convey an exhaustive listing of all possible sensors. Instead, it will be understood that these teachings will accommodate sensing any of a wide variety of circumstances or phenomena to support the operating functionality of the motorized transport unit 102 in a given application setting.


By one optional approach an audio input 416 (such as a microphone) and/or an audio output 418 (such as a speaker) can also operably couple to the control circuit 406. So configured the control circuit 406 can provide a variety of audible sounds to thereby communicate with a user of the motorized transport unit 102, other persons in the vicinity of the motorized transport unit 102, or even other motorized transport units 102 in the area. These audible sounds can include any of a variety of tones and other non-verbal sounds. These audible sounds can also include, in lieu of the foregoing or in combination therewith, pre-recorded or synthesized speech.


The audio input 416, in turn, provides a mechanism whereby, for example, a user provides verbal input to the control circuit 406. That verbal input can comprise, for example, instructions, inquiries, or information. So configured, a user can provide, for example, a question to the motorized transport unit 102 (such as, “Where are the towels?”). The control circuit 406 can cause that verbalized question to be transmitted to the central computer system 106 via the motorized transport unit's wireless transceiver 412. The central computer system 106 can process that verbal input to recognize the speech content and to then determine an appropriate response. That response might comprise, for example, transmitting back to the motorized transport unit 102 specific instructions regarding how to move the motorized transport unit 102 (via the aforementioned motorized wheel system 410) to the location in the shopping facility 101 where the towels are displayed.


In this example the motorized transport unit 102 includes a rechargeable power source 420 such as one or more batteries. The power provided by the rechargeable power source 420 can be made available to whichever components of the motorized transport unit 102 require electrical energy. By one approach the motorized transport unit 102 includes a plug or other electrically conductive interface that the control circuit 406 can utilize to automatically connect to an external source of electrical energy to thereby recharge the rechargeable power source 420.


By one approach the motorized transport unit 102 comprises an integral part of a movable item container 104 such as a grocery cart. As used herein, this reference to “integral” will be understood to refer to a non-temporary combination and joinder that is sufficiently complete so as to consider the combined elements to be as one. Such a joinder can be facilitated in a number of ways including by securing the motorized transport unit housing 402 to the item container using bolts or other threaded fasteners as versus, for example, a clip.


These teachings will also accommodate selectively and temporarily attaching the motorized transport unit 102 to an item container 104. In such a case the motorized transport unit 102 can include a movable item container coupling structure 422. By one approach this movable item container coupling structure 422 operably couples to a control circuit 202 to thereby permit the latter to control, for example, the latched and unlatched states of the movable item container coupling structure 422. So configured, by one approach the control circuit 406 can automatically and selectively move the motorized transport unit 102 (via the motorized wheel system 410) towards a particular item container until the movable item container coupling structure 422 can engage the item container to thereby temporarily physically couple the motorized transport unit 102 to the item container. So latched, the motorized transport unit 102 can then cause the item container to move with the motorized transport unit 102.


In embodiments such as illustrated in FIGS. 2A-3B, the movable item container coupling structure 422 includes a lifting system (e.g., including an electric drive or motor) to cause a portion of the body or housing 402 to engage and lift a portion of the item container off of the ground such that the motorized transport unit 102 can carry a portion of the item container. In other embodiments, the movable transport unit latches to a portion of the movable item container without lifting a portion thereof off of the ground.


In either case, by combining the motorized transport unit 102 with an item container, and by controlling movement of the motorized transport unit 102 via the aforementioned central computer system 106, these teachings will facilitate a wide variety of useful ways to assist both customers and associates in a shopping facility setting. For example, the motorized transport unit 102 can be configured to follow a particular customer as they shop within the shopping facility 101. The customer can then place items they intend to purchase into the item container that is associated with the motorized transport unit 102.


In some embodiments, the motorized transport unit 102 includes an input/output (I/O) device 424 that is coupled to the control circuit 406. The I/O device 424 allows an external device to couple to the control unit 404. The function and purpose of connecting devices will depend on the application. In some examples, devices connecting to the I/O device 424 may add functionality to the control unit 404, allow the exporting of data from the control unit 404, allow the diagnosing of the motorized transport unit 102, and so on.


In some embodiments, the motorized transport unit 102 includes a user interface 426 including for example, user inputs and/or user outputs or displays depending on the intended interaction with the user. For example, user inputs could include any input device such as buttons, knobs, switches, touch sensitive surfaces or display screens, and so on. Example user outputs include lights, display screens, and so on. The user interface 426 may work together with or separate from any user interface implemented at a user interface unit 114 (such as a smart phone or tablet device).


The control unit 404 includes a memory 408 coupled to the control circuit 406 and that stores, for example, operating instructions and/or useful data. The control circuit 406 can comprise a fixed-purpose hard-wired platform or can comprise a partially or wholly programmable platform. These architectural options are well known and understood in the art and require no further description here.


This control circuit 406 is configured (for example, by using corresponding programming stored in the memory 408 as will be well understood by those skilled in the art) to carry out one or more of the steps, actions, and/or functions described herein. The memory 408 may be integral to the control circuit 406 or can be physically discrete (in whole or in part) from the control circuit 406 as desired. This memory 408 can also be local with respect to the control circuit 406 (where, for example, both share a common circuit board, chassis, power supply, and/or housing) or can be partially or wholly remote with respect to the control circuit 406. This memory 408 can serve, for example, to non-transitorily store the computer instructions that, when executed by the control circuit 406, cause the control circuit 406 to behave as described herein. (As used herein, this reference to “non-transitorily” will be understood to refer to a non-ephemeral state for the stored contents (and hence excludes when the stored contents merely constitute signals or waves) rather than volatility of the storage media itself and hence includes both non-volatile memory (such as read-only memory (ROM) as well as volatile memory (such as an erasable programmable read-only memory (EPROM).)


It is noted that not all components illustrated in FIG. 4 are included in all embodiments of the motorized transport unit 102. That is, some components may be optional depending on the implementation.



FIG. 5 illustrates a functional block diagram that may generally represent any number of various electronic components of the system 100 that are computer type devices. The computer device 500 includes a control circuit 502, a memory 504, a user interface 506 and an input/output (I/O) interface 508 providing any type of wired and/or wireless connectivity to the computer device 500, all coupled to a communication bus 510 to allow data and signaling to pass therebetween. Generally, the control circuit 502 and the memory 504 may be referred to as a control unit. The control circuit 502, the memory 504, the user interface 506 and the I/O interface 508 may be any of the devices described herein or as understood in the art. The functionality of the computer device 500 will depend on the programming stored in the memory 504. The computer device 500 may represent a high level diagram for one or more of the central computer system 106, the motorized transport unit 102, the user interface unit 114, the location detection system 116, the user interface computer 128, the MTU docking station 122 and the MTU dispenser 120, or any other device or component in the system that is implemented as a computer device.


Additional Features Overview


Referring generally to FIGS. 1-5, the shopping assistance system 100 may implement one or more of several different features depending on the configuration of the system and its components. The following provides a brief description of several additional features that could be implemented by the system. One or more of these features could also be implemented in other systems separate from embodiments of the system. This is not meant to be an exhaustive description of all features and not meant to be an exhaustive description of the details any one of the features. Further details with regards to one or more features beyond this overview may be provided herein.


Tagalong Steering: This feature allows a given motorized transport unit 102 to lead or follow a user (e.g., a customer and/or a worker) throughout the shopping facility 101. For example, the central computer system 106 uses the location detection system 116 to determine the location of the motorized transport unit 102. For example, LED smart lights (e.g., the ByteLight system) of the location detection system 116 transmit a location number to smart devices which are with the customer (e.g., user interface units 114), and/or on the item container 104/motorized transport unit 102. The central computer system 106 receives the LED location numbers received by the smart devices through the wireless network 124. Using this information, in some embodiments, the central computer system 106 uses a grid placed upon a 2D CAD map and 3D point cloud model (e.g., from the databases 126) to direct, track, and plot paths for the other devices. Using the grid, the motorized transport unit 102 can drive a movable item container 104 in a straight path rather than zigzagging around the facility. As the user moves from one grid to another, the motorized transport unit 102 drives the container 104 from one grid to the other. In some embodiments, as the user moves towards the motorized transport unit, it stays still until the customer moves beyond an adjoining grid.


Detecting Objects: In some embodiments, motorized transport units 102 detect objects through several sensors mounted on motorized transport unit 102, through independent cameras (e.g., video cameras 118), through sensors of a corresponding movable item container 104, and through communications with the central computer system 106. In some embodiments, with semi-autonomous capabilities, the motorized transport unit 102 will attempt to avoid obstacles, and if unable to avoid, it will notify the central computer system 106 of an exception condition. In some embodiments, using sensors 414 (such as distance measurement units, e.g., laser or other optical-based distance measurement sensors), the motorized transport unit 102 detects obstacles in its path, and will move to avoid, or stop until the obstacle is clear.


Visual Remote Steering: This feature enables movement and/or operation of a motorized transport unit 102 to be controlled by a user on-site, off-site, or anywhere in the world. This is due to the architecture of some embodiments where the central computer system 106 outputs the control signals to the motorized transport unit 102. These controls signals could have originated at any device in communication with the central computer system 106. For example, the movement signals sent to the motorized transport unit 102 may be movement instructions determined by the central computer system 106; commands received at a user interface unit 114 from a user; and commands received at the central computer system 106 from a remote user not located at the shopping facility space.


Determining Location: Similar to that described above, this feature enables the central computer system 106 to determine the location of devices in the shopping facility 101. For example, the central computer system 106 maps received LED light transmissions, Bluetooth low energy radio signals or audio signals (or other received signals encoded with location data) to a 2D map of the shopping facility. Objects within the area of the shopping facility are also mapped and associated with those transmissions. Using this information, the central computer system 106 can determine the location of devices such as motorized transport units.


Digital Physical Map Integration: In some embodiments, the system 100 is capable of integrating 2D and 3D maps of the shopping facility with physical locations of objects and workers. Once the central computer system 106 maps all objects to specific locations using algorithms, measurements and LED geo-location, for example, grids are applied which sections off the maps into access ways and blocked sections. Motorized transport units 102 use these grids for navigation and recognition. In some cases, grids are applied to 2D horizontal maps along with 3D models. In some cases, grids start at a higher unit level and then can be broken down into smaller units of measure by the central computer system 106 when needed to provide more accuracy.


Calling a Motorized Transport Unit: This feature provides multiple methods to request and schedule a motorized transport unit 102 for assistance in the shopping facility. In some embodiments, users can request use of a motorized transport unit 102 through the user interface unit 114. The central computer system 106 can check to see if there is an available motorized transport unit. Once assigned to a given user, other users will not be able to control the already assigned transport unit. Workers, such as store associates, may also reserve multiple motorized transport units in order to accomplish a coordinated large job.


Locker Delivery: In some embodiments, one or more motorized transport units 102 may be used to pick, pack, and deliver items to a particular storage locker 132. The motorized transport units 102 can couple to and move the storage locker to a desired location. In some embodiments, once delivered, the requestor will be notified that the items are ready to be picked up, and will be provided the locker location and locker security code key.


Route Optimization: In some embodiments, the central computer system automatically generates a travel route for one or more motorized transport units through the shopping facility space. In some embodiments, this route is based on one or more of a user provided list of items entered by the user via a user interface unit 114; user selected route preferences entered by the user via the user interface unit 114; user profile data received from a user information database (e.g., from one of databases 126); and product availability information from a retail inventory database (e.g., from one of databases 126). In some cases, the route intends to minimize the time it takes to get through the facility, and in some cases, may route the shopper to the least busy checkout area. Frequently, there will be multiple possible optimum routes. The route chosen may take the user by things the user is more likely to purchase (in case they forgot something), and away from things they are not likely to buy (to avoid embarrassment). That is, routing a customer through sporting goods, women's lingerie, baby food, or feminine products, who has never purchased such products based on past customer behavior would be non-productive, and potentially embarrassing to the customer. In some cases, a route may be determined from multiple possible routes based on past shopping behavior, e.g., if the customer typically buys a cold Diet Coke product, children's shoes or power tools, this information would be used to add weight to the best alternative routes, and determine the route accordingly.


Store Facing Features: In some embodiments, these features enable functions to support workers in performing store functions. For example, the system can assist workers to know what products and items are on the shelves and which ones need attention. For example, using 3D scanning and point cloud measurements, the central computer system can determine where products are supposed to be, enabling workers to be alerted to facing or zoning of issues along with potential inventory issues.


Phone Home: This feature allows users in a shopping facility 101 to be able to contact remote users who are not at the shopping facility 101 and include them in the shopping experience. For example, the user interface unit 114 may allow the user to place a voice call, a video call, or send a text message. With video call capabilities, a remote person can virtually accompany an in-store shopper, visually sharing the shopping experience while seeing and talking with the shopper. One or more remote shoppers may join the experience.


Returns: In some embodiments, the central computer system 106 can task a motorized transport unit 102 to keep the returns area clear of returned merchandise. For example, the transport unit may be instructed to move a cart from the returns area to a different department or area. Such commands may be initiated from video analytics (the central computer system analyzing camera footage showing a cart full), from an associate command (digital or verbal), or on a schedule, as other priority tasks allow. The motorized transport unit 102 can first bring an empty cart to the returns area, prior to removing a full one.


Bring a Container: One or more motorized transport units can retrieve a movable item container 104 (such as a shopping cart) to use. For example, upon a customer or worker request, the motorized transport unit 102 can re-position one or more item containers 104 from one location to another. In some cases, the system instructs the motorized transport unit where to obtain an empty item container for use. For example, the system can recognize an empty and idle item container that has been abandoned or instruct that one be retrieved from a cart storage area. In some cases, the call to retrieve an item container may be initiated through a call button placed throughout the facility, or through the interface of a user interface unit 114.


Respond to Voice Commands: In some cases, control of a given motorized transport unit is implemented through the acceptance of voice commands. For example, the user may speak voice commands to the motorized transport unit 102 itself and/or to the user interface unit 114. In some embodiments, a voice print is used to authorize to use of a motorized transport unit 102 to allow voice commands from single user at a time.


Retrieve Abandoned Item Containers: This feature allows the central computer system to track movement of movable item containers in and around the area of the shopping facility 101, including both the sale floor areas and the back-room areas. For example, using visual recognition through store cameras 118 or through user interface units 114, the central computer system 106 can identify abandoned and out-of-place movable item containers. In some cases, each movable item container has a transmitter or smart device which will send a unique identifier to facilitate tracking or other tasks and its position using LED geo-location identification. Using LED geo-location identification with the Determining Location feature through smart devices on each cart, the central computer system 106 can determine the length of time a movable item container 104 is stationary.


Stocker Assistance: This feature allows the central computer system to track movement of merchandise flow into and around the back-room areas. For example, using visual recognition and captured images, the central computer system 106 can determine if carts are loaded or not for moving merchandise between the back room areas and the sale floor areas. Tasks or alerts may be sent to workers to assign tasks.


Self-Docking: Motorized transport units 102 will run low or out of power when used. Before this happens, the motorized transport units 102 need to recharge to stay in service. According to this feature, motorized transport units 102 will self-dock and recharge (e.g., at a MTU docking station 122) to stay at maximum efficiency, when not in use. When use is completed, the motorized transport unit 102 will return to a docking station 122. In some cases, if the power is running low during use, a replacement motorized transport unit can be assigned to move into position and replace the motorized transport unit with low power. The transition from one unit to the next can be seamless to the user.


Item Container Retrieval: With this feature, the central computer system 106 can cause multiple motorized transport units 102 to retrieve abandoned item containers from exterior areas such as parking lots. For example, multiple motorized transport units are loaded into a movable dispenser, e.g., the motorized transport units are vertically stacked in the dispenser. The dispenser is moved to the exterior area and the transport units are dispensed. Based on video analytics, it is determined which item containers 104 are abandoned and for how long. A transport unit will attach to an abandoned cart and return it to a storage bay.


Motorized Transport Unit Dispenser: This feature provides the movable dispenser that contains and moves a group of motorized transport units to a given area (e.g., an exterior area such as a parking lot) to be dispensed for use. For example, motorized transport units can be moved to the parking lot to retrieve abandoned item containers 104. In some cases, the interior of the dispenser includes helically wound guide rails that mate with the guide member 208 to allow the motorized transport units to be guided to a position to be dispensed.


Specialized Module Retrieval: This feature allows the system 100 to track movement of merchandise flow into and around the sales floor areas and the back-room areas including special modules that may be needed to move to the sales floor. For example, using video analytics, the system can determine if a modular unit it loaded or empty. Such modular units may house items that are of seasonal or temporary use on the sales floor. For example, when it is raining, it is useful to move a module unit displaying umbrellas from a back room area (or a lesser accessed area of the sales floor) to a desired area of the sales floor area.


Authentication: By one approach this feature uses a voice imprint with an authentication code/word to authenticate a user to a given motorized transport unit. If desired, one motorized transport unit can be swapped for another using this authentication. For example, a token may be used during the session with the user. The token can be a unique identifier for the session that is dropped once the session ends. A logical token may be a session identifier (ID) used by the application of the user interface unit 114 to establish the session ID when a user logs on and when deciding to so use the system 100. In some embodiments, communications throughout the session are encrypted using the Secure Sockets Layer (SSL) standard or other methods at the (or a) transport level.


Further Details of Some Embodiments


In accordance with some embodiments, further details are now provided for one or more of these and other features.



FIG. 6 presents a process 600 that can be carried out by the aforementioned central computer system 106 to assign the aforementioned motorized transport unit 102 to a requesting user to serve as a motorized mobile retail facility personal assistance apparatus. In this illustrative example the central computer system 106 receives (via, for example, a wireless communication from a user interface unit 114 as described above) a request 601 from a user to use one of an available plurality of motorized mobile retail facility personal assistance apparatuses.


By one approach, and as illustrated in FIG. 6, the central computer system 106 assesses 602 that user request 601. This assessment can include, for example, determining that the user request 601 is properly formatted and includes all relevant information (including, for example, information to identify the user).


This assessment can also optionally include, if desired, accessing other information 603 such as information indicating whether the user is personally barred from being assigned a motorized mobile retail facility personal assistance apparatus and/or information regarding a present physical location of the user. A particular user may be personally barred based upon prior misuse of a motorized mobile retail facility personal assistance apparatus or other misconduct in a shopping facility 101 such as a retail store environment. The present physical location of the user may be useful to determine, for example, that a user making a present request is not actually at the retail store and hence is not physically present to accept the requested assignment.


When the assessment 602 concludes with a determination that the user request 601 should be denied, this process 600 can provide for such a denial 604. That denial state can then be transmitted to the user interface unit 114, transmitted to a relevant person of authority at the shopping facility 101, entered into a log, and so forth as desired.


When the assessment 602 concludes with a determination that the user request 601 can be granted, this process 600 provides for assigning 605 an available motorized mobile retail facility personal assistance apparatus. This assignment can include, for example, causing the assigned apparatus to maneuver and otherwise make itself available to the requesting user.


By one approach, assigning the available apparatus includes wirelessly transmitting to the user via the user interface unit 114 a globally unique session token 606 that uniquely corresponds to this usage session. The globally unique session token can comprise an appropriate string of characters of choice (including alphabetic characters, a numeric character, punctuation characters, hexadecimal characters, and so forth as desired). As used herein, this reference to being “globally unique” means that the session token is an utterly unique string of characters within the context of session tokens utilized by the entity that owns and/or operates the shopping facility 101 to thereby uniquely identify each usage session with such session tokens.


By another approach, in lieu of the foregoing or in combination therewith, assigning the available apparatus includes wirelessly transmitting to the user via the user interface unit 114 a temporary N-digit code 607. For many application settings it will be useful for “N” to be a non-zero integer. In this particular illustrative example “N” has at least five digits. In such a case, these teachings contemplate that the user interface unit 114 will display the N-digit code to the user for no more than a pre-determined period of time (such as 10 seconds, 30 seconds, or some other duration of choice). The central computer system 106 then monitors to determine whether a communication including the N-digit code is received 608 within a pre-determined period of time “T.”


When the N-digit code is not so received, the central computer system 106 can withdraw the assignment 609 of the motorized mobile retail facility personal assistance apparatus. Withdrawing the assignment 609 can include, for example, instructing the motorized mobile retail facility personal assistance apparatus to return to a waiting/staging area and to retire the globally unique session token that had been assigned to the granted usage session.


In addition to the foregoing, or in lieu thereof, these teachings will also accommodate having the central computer system 106 track 610 information regarding use of the motorized mobile retail facility personal assistance apparatus. These teachings will accommodate a variety of usage information including information regarding activity of the motorized mobile retail facility personal assistance apparatus as well as activity of the user. This activity information can include, for example, physical location information regarding these respective entities. By one approach the aforementioned activity information can include a timestamp or other temporal indicator to correlate the tracked activities with specific times.


By one approach the central computer system 106 is configured to use the aforementioned globally unique session token to log some or all of the aforementioned tracking information. So configured, such activity information can be readily recalled or otherwise processed as a function of that globally unique session token to thereby correlate that tracked activity with a specific usage session.


By one optional approach, and as illustrated at decision block 611, this process 600 will accommodate determining whether the user (and/or the motorized mobile retail facility personal assistance apparatus) enters an area that is off-limits to the assigned motorized mobile retail facility personal assistance apparatus. An example of an off-limits area might be a back stockroom where only authorized associates of the shopping facility 101 are permitted. Another example might be any area that is external to the shopping facility 101 itself (in which case the concept of “off limits” might pertain to the assigned motorized mobile retail facility personal assistance apparatus though not to the user themselves).


Upon detecting an off-limits incursion, the central computer system 106 can respond appropriately. In this particular example the central computer system 106 withdraws the assignment 609 of the assigned motorized mobile retail facility personal assistance apparatus and thereby concludes the usage session as described above. These teachings will accommodate other approaches if desired. For example, the assigned motorized mobile retail facility personal assistance apparatus can stop moving in conjunction with the user and can sound a caution or alert tone or message in an attempt to redirect the user from the off-limits area.


It is possible that an assigned motorized mobile retail facility personal assistance apparatus may need replacement during a usage session. As one simple example in these regards the rechargeable power source 420 for the apparatus 102 may require recharging. As another simple example the assigned apparatus may suffer a breakdown or other fault that prevents or impairs the ability of the assigned apparatus to itself conclude the usage session in a satisfactory manner. In such a case, if desired, this process 600 can be optionally configured to determine (at decision block 612) a need to replace an assigned motorized mobile retail facility personal assistance apparatus, and upon detecting such a need, reassign 613 a different motorized mobile retail facility personal assistance apparatus to the user for that usage session.


This reassignment 613 can include, for example, transferring the globally unique session token that was assigned to the user for this usage session to the replacement motorized mobile retail facility personal assistance apparatus. Using this approach avoids any need to update the user interface unit 114 with a new globally unique session token and also ensures continuity when tracking information during a particular usage session as described above.


These teachings will also include detecting or determining (at block 614) an indication from the user (via the corresponding user interface unit 114) that the user wishes to conclude the usage session. Upon detecting/determining that input, the central computer system 106 can end the session 615. Ending the session can include, for example, returning the assigned motorized mobile retail facility personal assistance apparatus to a waiting area to await a new assignment, reassigning the apparatus to a new usage session, directing the apparatus to a recharging location, or such other action as may be appropriate to the circumstances and application setting as desired.


When configured as described in the above example, it can be helpful for the user interface unit 114 to respond in an expected manner to the central computer system 106. FIG. 7 presents a process 700 that can be carried out by the user interface unit 114 in these regards.


In this example the user interface unit 114 receives (at block 701) a request from a user 702 of the user interface unit 114 for a motorized mobile retail facility personal assistance apparatus. This request might be entered by the user 702 using any of a variety of selection or data-entry methodologies as are known in the art. As the present teachings are not overly sensitive to any particular choices in these regards, further elaboration in these regards is not provided here for the sake of brevity.


Upon receiving this request the user interface unit 114 transmits that request at block 703. As noted above, this transmission can comprise a wireless transmission to the central computer system 106 using, for example, an intervening Wi-Fi network or some other wireless technology of choice.


As noted above, in some configurations the central computer system 106 may deny the user's request. In that case the central computer system 106 may transmit to the user interface unit 114 a corresponding denial message. Upon detecting receipt of such a denial at block 704, this process 700 will accommodate displaying that denial at block 705 using, for example, an integral display.


In any event, presuming that the central computer system 106 grants the aforementioned request, the user interface unit 114 receives (at block 706) the above-described globally unique session token. In an application setting where the central computer system 106 also transmits the above-described temporary N-digit code, at block 707 the user interface unit 114 will also receive that temporary N-digit code.


At block 708 the user interface unit 114 will use that received globally unique session token as an encryption key for session communications as correspond to this particular usage session for the assigned motorized mobile retail facility personal assistance apparatus. Accordingly, that globally unique session token may serve as an encryption key for communications (both transmitted and received) between the user interface unit 114 and the assigned motorized mobile retail facility personal assistance apparatus and/or for communications (both transmitted and received) between the user interface unit 114 and the central computer system 106. If desired, that globally unique session token may also serve, during (and only during) the usage session as an encryption key for communications (both transmitted and received) between the assigned motorized mobile retail facility personal assistance apparatus and the central computer system 106.


In the case when the user interface unit 114 also receives the temporary N-digit code, at block 709 the user interface unit 114 can display that temporary N-digit code using, for example, its integral display. FIG. 8 provides an illustrative example in these regards. In this example the user interface unit 114 includes a display that comprises a touch-sensitive display as is known in the art. The received N-digit code in this example comprises a five-digit code 802 comprising the digits “83835” which are displayed on the user interface unit's display.


With continued reference to both FIGS. 7 and 8, at block 710 the user interface unit 114 determines whether the user 702 enters the N-digit code into the user interface unit 114 within a predetermined period of time (such as 10 seconds, 30 seconds, or some other duration of choice). As illustrated in FIG. 8, the user interface unit 114 can present a code-entry field 803 to accommodate entry of the N-digit code on a digit-by-digit basis. In this particular example the user 702 has entered the first two digits (i.e., “8” and “3”) of the five-digit code 802 and accordingly the code-entry insertion point 806 is displayed at the third code-entry opportunity in the code-entry field 803.


In this example the user interface unit 114 also displays a numeric keypad 804 that the user 702 can employ to enter the N-digit code. Upon completing entry of the code, the user 702 then asserts an enter button 805 that the user interface unit 114 also presents on the display.


At block 711 the user interface unit 114 then transmits the entered N-digit code to the central computer system 106 for processing as described above. By one approach this transmission is encrypted using the already-assigned globally unique session token.



FIGS. 9 through 11 present further illustrative examples in these same general regards. In particular, FIG. 9 illustrates a process 900 that can serve in an application setting as shown in FIG. 10 where the retail shopping facility 101 includes a plurality of motorized transport units 102 where at least some of the motorized transport units 102 are differently purposed and hence differently physically configured as compared with other. FIG. 10 provides a number of specific examples in these regards. It will be understood that the specific details in this example are intended to serve an illustrative purpose and are not intended to suggest any particular limitations with respect to these teachings.


In particular, in this example, the plurality of available motorized transport units 102 include one or more disability mobility platforms 1001 (purposed and configured to support and convey, for example, a person having a walking disability), one or more grocery carts 1002 (purposed and configured to contain and convey, for example, one or more grocery items in a suitably-sized basket), one or more pallets 1003 (purposed and configured, for example, as a largely or even completely flat surface (lacking side walls) that can support thereon one or more items to be moved from one place to another), one or more child-carrying platforms 1004 (purposed and configured to support and convey, for example, one or more infants in a seated posture or in a child carrier), and any other platform 1005 or the like specifically configured to accommodate two or more of the foregoing purposes and/or to support any other desired capability.


These motorized transport units 102 can be made available, for example, in a partially or wholly dedicated holding or staging area inside the retail shopping facility 101 or nearby the retail shopping facility 101 as desired.


With continued reference to both FIGS. 9 and 10, pursuant to this process 900 the aforementioned central computer system 106 (at block 901) interfaces with a user 1006 regarding tasking a particular one of plurality of motorized transport units 102 with assisting the user 1006 in a particular corresponding way at the retail shopping facility 101. The central computer system 106 can effect this interfacing via its network interface(s) 112 as described above and via one or more intervening networks 124, also as described above.


The user 1006, in turn, can interface with the central computer system 106 via, for example, the aforementioned user interface unit 114. These teachings are highly flexible in these regards and will accommodate, for example, having the central computer system 106 so interface with the user 1006 via one or more of a web-based interface, a short-range wireless interface, a voice-based interface, and/or a gesture-based interface. All of these interfaces are well known in the art. Since the present teachings are not especially sensitive to any particular choices in these regards, further elaboration is not provided here for the sake of brevity.



FIG. 11 provides some specific examples regarding the aforementioned interfacing. By one approach, and as illustrated at block 1101, the central computer system 106 interfaces with the user 1006 by prompting the user to identify a particular purpose (or purposes) and/or configuration (or configurations) for the motorized transport unit 102. The supplied information can then be utilized by the central computer system 106 to determine, for example, how many and what kind of independent purposes the user seeks to have served by the motorized transport unit(s) 102. As one example in these regards, the user 1006 may respond to such a prompting by identifying their needs as including shopping for groceries while accompanied by one infant.


These teachings will accommodate a variety of ways of eliciting such information. By one approach, for example, the user 1006 can be prompted to identify a particular response by selecting from amongst a plurality of provided candidate responses. By another approach the user 1006 can respond using a free-text field or by otherwise responding in a natural sentence-based way with a description of their needs.


By another approach, in lieu of the foregoing or in combination therewith, the process can prompt the user 1006 to determine how many of the motorized transport units to simultaneously task with assisting the user 1006 at the retail shopping facility 101. For example, a user 1006 may have simultaneous need of both a grocery cart 1002 (to gather a plurality of ordinary grocery items) and a pallet 1003 (to acquire, for example, a particularly large item such as a flatscreen television).


By yet another approach, and again in lieu of the foregoing or in combination therewith, the process can prompt the user 1006 to identify a time at which the assistance of the assigned motorized transport unit 102 is sought. This can comprise identifying a specific day and time, or, for example, a window of time (for example, from 3:30 PM to 4 PM on a particular day).


Referring again to FIG. 9, and based upon the foregoing interfacing active, by one optional approach the central computer system (at optional block 902) schedules and/or assigns at least one appropriate motorized transport unit 102 that will meet the specified needs of the user 1006. For example, if the user 1006 intends to shop for groceries while accompanied by an infant, the central computer system 106 may assign a motorized transport unit 102 that includes both a cart for receiving groceries as well as a seat to safely receive and contain the infant. Or, when an integrated platform is not available that can support both of those needs, the central computer system can assign a grocery cart 1002 and a separate child-carrying platform 1004 to assist this user 1006 during this visit to the retail shopping facility 101.


So configured, using a combination of automation, one or more remote resources, and the appropriate interactions of the user themselves a motorized mobile retail facility personal assistance apparatus can be reliably and timely assigned to a given individual in a short period of time and with little or no training required on the part of the user.


In some embodiments, systems, apparatuses and methods are provided herein useful to provide customer assistance. In some embodiments, In some embodiments, an apparatus comprises: a plurality of motorized transport units configured to move through a retail shopping facility, as least some of the motorized transport units being differently purposed and hence differently physically configured as compared to one another; a central computer system having a network interface such that the central computer system wirelessly communicates with the plurality of motorized transport units and wherein the central computer system is configured to: interface with a user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility.


In some embodiments, the motorized transport units include at least one unit that is purposed and physically configured to serve as one of: a disability mobility platform; a grocery cart; a pallet; a child-carrying platform. In some embodiments, the motorized transport units include at least one other unit that is purposed and physically configured differently than the at least one unit and which is purposed and physically configured to serve as one of: a disability mobility platform; a grocery cart; a pallet; a child-carrying platform. In some embodiments, the central computer system is configured to interface with the user via at least one of: a web-based interface; a short-range wireless interface; a voice-based interface; a gesture-based interface. In some embodiments, the central computer system is configured to interface with the user via any of: a web-based interface; a short-range wireless interface; a voice-based interface; a gesture-based interface. In some embodiments, the central computer system is configured to interface with the user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility by, at least in part, prompting the user to identify a particular purpose or configuration for the motorized transport unit to be tasked with assisting the user. In some embodiments, the central computer system is further configured to interface with the user to determine how many of the motorized transport units to simultaneously task with assisting the user at the retail shopping facility. In some embodiments, the central computer system is configured to determine how many of the motorized transport units to simultaneously task with assisting the user at the retail shopping facility, at least in part, by determining how many independent purposes the user seeks to be served. In some embodiments, the central computer system is configured to interface with the user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility by, at least in part, prompting the user to identify a time at which the assistance is sought. In some embodiments, the central computer system is configured to interface with the user to schedule the time at which at least one of the motorized transport units will be available at the retail shopping facility to assist the user.


In some embodiments, a method comprises: by a central computer system having a network interface such that the central computer system wirelessly communicates with a plurality of motorized transport units that are configured to move through a retail shopping facility, wherein as least some of the motorized transport units are differently purposed and hence differently physically configured as compared to one another: interfacing with a user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility.


In some embodiments, the motorized transport units include at least one unit that is purposed and physically configured to serve as one of: a disability mobility platform; a grocery cart; a pallet; a child-carrying platform. In some embodiments, the motorized transport units include at least one other unit that is purposed and physically configured differently than the at least one unit and which is purposed and physically configured to serve as one of: a disability mobility platform; a grocery cart; a pallet; a child-carrying platform. In some embodiments, the central computer system is configured to interface with the user via at least one of: a web-based interface; a short-range wireless interface; a voice-based interface; a gesture-based interface. In some embodiments, the central computer system is configured to interface with the user via any of: a web-based interface; a short-range wireless interface; a voice-based interface; a gesture-based interface. In some embodiments, the interfacing with the user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility comprises, at least in part, prompting the user to identify a particular purpose or configuration for the motorized transport unit to be tasked with assisting the user. In some embodiments, the interfacing with the user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility further comprises interfacing with the user to determine how many of the motorized transport units to simultaneously task with assisting the user at the retail shopping facility. In some embodiments, the determining how many of the motorized transport units to simultaneously task with assisting the user at the retail shopping facility comprises, at least in part, determining how many independent purposes the user seeks to be served. In some embodiments, the interfacing with the user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility comprises, at least in part, prompting the user to identify a time at which the assistance is sought. In some embodiments, the interfacing with the user regarding tasking a particular one of the plurality of motorized transport units with assisting the user in a particular corresponding way at the retail shopping facility further comprises interfacing with the user to schedule the time at which at least one of the motorized transport units will be available at the retail shopping facility to assist the user.


Those skilled in the art will recognize that a wide variety of modifications, alterations, and combinations can be made with respect to the above described embodiments without departing from the scope of the invention, and that such modifications, alterations, and combinations are to be viewed as being within the ambit of the inventive concept.

Claims
  • 1. An apparatus comprising: a plurality of motorized mobile retail facility personal assistance apparatuses that are each configured to move through a retail shopping facility;a central computer system having a network interface such that the central computer system wirelessly communicates with the plurality of motorized mobile retail facility personal assistance apparatuses and with at least one user interface unit, the central computer system being configured to: receive a request from a user via the at least one user interface unit to use one of the plurality of motorized mobile retail facility personal assistance apparatuses;assign an available one of the plurality of motorized mobile retail facility personal assistance apparatuses for use by the user during a usage session by, at least in part, wirelessly transmitting to the user via the user interface unit a globally unique session token that uniquely corresponds to the usage session and also by wirelessly transmitting to the user interface unit a temporary N-digit code (where “N” is a non-zero integer) to be displayed by the user interface unit for no more than a pre-determined period of time, wherein beginning the usage session occurs when the user enters the temporary N-digit code into the user interface unit before expiration of the pre-determined period of time.
  • 2. The apparatus of claim 1 wherein at least some of the plurality of motorized mobile retail facility personal assistance apparatuses are motorized transport devices that automatically connect, when assigned by the central computer system, to a movable item container such that the motorized transport device will move the movable item container to a location of convenience as corresponds to the user during the usage session such that the user can readily place a selected item into the movable item container to facilitate purchasing the selected item.
  • 3. The apparatus of claim 1 wherein the user interface unit comprises at least one of a smartphone and a pad/tablet-styled computer.
  • 4. The apparatus of claim 3 wherein the user interface unit comprises one of: a movable item container-mounted two-way communications unit; anda user-possessed personally-portable two-way communications unit.
  • 5. The apparatus of claim 1 wherein the motorized mobile retail facility personal assistance apparatus is configured to use the globally unique session token as an encryption key when communicating with the user interface unit during at least part of the usage session.
  • 6. The apparatus of claim 1 wherein N has at least five digits.
  • 7. The apparatus of claim 1 further comprising retiring the globally unique session token when the user fails to enter the temporary N-digit code into the user interface unit before expiration of the pre-determined period of time and hence withdrawing the assignment of the motorized mobile retail facility personal assistance apparatus for use by the user during a usage session.
  • 8. The apparatus of claim 1 wherein the central computer system is further configured to use the globally unique session token to log tracking information of the user's use of the motorized mobile retail facility personal assistance apparatus during the usage session.
  • 9. The apparatus of claim 8 wherein the tracking information includes information regarding activity of the motorized mobile retail facility personal assistance apparatus and activity of the user.
  • 10. The apparatus of claim 1 wherein the central computer system is further configured to assess the request to use one of the plurality of motorized mobile retail facility personal assistance apparatuses before assigning an available one of the plurality of motorized mobile retail facility personal assistance apparatuses for use by the user during a usage session.
  • 11. The apparatus of claim 1 wherein the central computer system is configured to assess the request by, at least in part, determining whether the user is personally barred from being assigned one of the motorized mobile retail facility personal assistance apparatuses.
  • 12. The apparatus of claim 1 wherein the central computer system is configured to assess the request by, at least in part, using a present physical location of the user.
  • 13. The apparatus of claim 1 wherein the central computer system is further configured to determine that the assigned motorized mobile retail facility personal assistance apparatus must be replaced during the usage session and arranging to transfer, during the usage session, the assignment to a replacement motorized mobile retail facility personal assistance apparatus.
  • 14. The apparatus of claim 13 wherein the central computer system is configured to arrange the transfer, at least in part, by transferring the globally unique session token to the replacement motorized mobile retail facility personal assistance apparatus.
  • 15. The apparatus of claim 1 wherein the assigned motorized mobile retail facility personal assistance apparatus provides a visual indication that corresponds to at least initiation of the usage session.
  • 16. The apparatus of claim 15 wherein the visual indication comprises an intermittent light.
  • 17. The apparatus of claim 1 wherein the central computer system is configured to terminate the usage session when the user enters an area that is off limits to the assigned motorized mobile retail facility personal assistance apparatus.
  • 18. The apparatus of claim 1 wherein the central computer system is configured to end the usage session upon receiving an indication from the user via the user interface unit that the user wishes to conclude the usage session.
RELATED APPLICATIONS

This application claims the benefit of each of the following U.S. Provisional applications, each of which is incorporated herein by reference in its entirety: U.S. Provisional Application No. 62/129,726, filed Mar. 6, 2015, U.S. Provisional Application No. 62/129,727, filed Mar. 6, 2015, U.S. Provisional Application No. 62/138,877, filed Mar. 26, 2015, U.S. Provisional Application No. 62/138,885, filed Mar. 26, 2015, U.S. Provisional Application No. 62/152,421, filed Apr. 24, 2015, U.S. Provisional Application No. 62/152,465, filed Apr. 24, 2015, U.S. Provisional Application No. 62/152,440, filed Apr. 24, 2015, U.S. Provisional Application No. 62/152,630, filed Apr. 24, 2015, U.S. Provisional Application No. 62/152,711, filed Apr. 24, 2015, U.S. Provisional Application No. 62/152,610, filed Apr. 24, 2015, U.S. Provisional Application No. 62/152,667, filed Apr. 24, 2015, U.S. Provisional Application No. 62/157,388, filed May 5, 2015, U.S. Provisional Application No. 62/165,579, filed May 22, 2015, U.S. Provisional Application No. 62/165,416, filed May 22, 2015, U.S. Provisional Application No. 62/165,586, filed May 22, 2015, U.S. Provisional Application No. 62/171,822, filed Jun. 5, 2015, U.S. Provisional Application No. 62/175,182, filed Jun. 12, 2015, U.S. Provisional Application No. 62/182,339, filed Jun. 19, 2015, U.S. Provisional Application No. 62/185,478, filed Jun. 26, 2015, U.S. Provisional Application No. 62/194,131, filed Jul. 17, 2015, U.S. Provisional Application No. 62/194,119, filed Jul. 17, 2015, U.S. Provisional Application No. 62/194,121, filed Jul. 17, 2015, U.S. Provisional Application No. 62/194,127, filed Jul. 17, 2015, U.S. Provisional Application No. 62/202,744, filed Aug. 7, 2015, U.S. Provisional Application No. 62/202,747, filed Aug. 7, 2015, U.S. Provisional Application No. 62/205,548, filed Aug. 14, 2015, U.S. Provisional Application No. 62/205,569, filed Aug. 14, 2015, U.S. Provisional Application No. 62/205,555, filed Aug. 14, 2015, U.S. Provisional Application No. 62/205,539, filed Aug. 14, 2015, U.S. Provisional Application No. 62/207,858, filed Aug. 20, 2015, U.S. Provisional Application No. 62/214,826, filed Sep. 4, 2015, U.S. Provisional Application No. 62/214,824, filed Sep. 4, 2015, U.S. Provisional Application No. 62/292,084, filed Feb. 5, 2016, U.S. Provisional Application No. 62/302,547, filed Mar. 2, 2016, U.S. Provisional Application No. 62/302,567, filed Mar. 2, 2016, U.S. Provisional Application No. 62/302,713, filed Mar. 2, 2016, and U.S. Provisional Application No. 62/303,021, filed Mar. 3, 2016.

US Referenced Citations (615)
Number Name Date Kind
1774653 Marriott Sep 1930 A
2669345 Brown Feb 1954 A
3765546 Westerling Oct 1973 A
4071740 Gogulski Jan 1978 A
4158416 Podesta Jun 1979 A
4588349 Reuter May 1986 A
4672280 Honjo Jun 1987 A
4777416 George, II Oct 1988 A
4791482 Barry Dec 1988 A
4868544 Havens Sep 1989 A
4911608 Krappitz Mar 1990 A
5119087 Lucas Jun 1992 A
5279672 Betker Jan 1994 A
5287266 Malec Feb 1994 A
5295551 Sukonick Mar 1994 A
5363305 Cox Nov 1994 A
5380138 Kasai Jan 1995 A
5384450 Goetz, Jr. Jan 1995 A
5395206 Cerny, Jr. Mar 1995 A
5402051 Fujiwara Mar 1995 A
5548515 Pilley Aug 1996 A
5632381 Thust May 1997 A
5652489 Kawakami Jul 1997 A
5671362 Cowe Sep 1997 A
5777571 Chuang Jul 1998 A
5801340 Peter Sep 1998 A
5917174 Moore Jun 1999 A
5920261 Hughes Jul 1999 A
5969317 Espy Oct 1999 A
6018397 Cloutier Jan 2000 A
6199753 Tracy Mar 2001 B1
6201203 Tilles Mar 2001 B1
6240342 Fiegert May 2001 B1
6339735 Peless Jan 2002 B1
6365857 Maehata Apr 2002 B1
6374155 Wallach Apr 2002 B1
6394519 Byers May 2002 B1
6431078 Serrano Aug 2002 B2
6522952 Arai Feb 2003 B1
6525509 Petersson Feb 2003 B1
6535793 Allard Mar 2003 B2
6550672 Tracy Apr 2003 B1
6571693 Kaldenberg Jun 2003 B1
6584375 Bancroft Jun 2003 B2
6584376 VanKommer Jun 2003 B1
6600418 Francis Jul 2003 B2
6601759 Fife Aug 2003 B2
6606411 Loui Aug 2003 B1
6626632 Guenzi Sep 2003 B2
6633800 Ward Oct 2003 B1
6655897 Harwell Dec 2003 B1
6667592 Jacobs Dec 2003 B2
6672601 Hofheins Jan 2004 B1
6678583 Nasr Jan 2004 B2
6688435 Will Feb 2004 B1
6728597 Didriksen Apr 2004 B2
6731204 Lehmann May 2004 B2
6745186 Testa Jun 2004 B1
6752582 Garcia Jun 2004 B2
6810149 Squilla Oct 2004 B1
6816085 Haynes Nov 2004 B1
6832884 Robinson Dec 2004 B2
6841963 Song Jan 2005 B2
6883201 Jones Apr 2005 B2
6885736 Uppaluru Apr 2005 B2
6886101 Glazer Apr 2005 B2
6895301 Mountz May 2005 B2
6910828 Hughes Jun 2005 B1
6937989 McIntyre Aug 2005 B2
6954695 Bonilla Oct 2005 B2
6967455 Nakadai Nov 2005 B2
6975997 Murakami Dec 2005 B1
7039499 Nasr May 2006 B1
7066291 Martins Jun 2006 B2
7101113 Hughes Sep 2006 B2
7101139 Benedict Sep 2006 B1
7117902 Osborne Oct 2006 B2
7145562 Schechter Dec 2006 B2
7147154 Myers Dec 2006 B2
7177820 McIntyre Feb 2007 B2
7184586 Jeon Feb 2007 B2
7205016 Garwood Apr 2007 B2
7206753 Bancroft Apr 2007 B2
7222363 Rice May 2007 B2
7233241 Overhultz Jun 2007 B2
7234609 DeLazzer Jun 2007 B2
7261511 Felder Aug 2007 B2
7367245 Okazaki May 2008 B2
7381022 King Jun 2008 B1
7402018 Mountz Jul 2008 B2
7431208 Feldman Oct 2008 B2
7447564 Yasukawa Nov 2008 B2
7463147 Laffoon Dec 2008 B1
7474945 Matsunaga Jan 2009 B2
7487913 Adema Feb 2009 B2
7533029 Mallett May 2009 B2
7554282 Nakamoto Jun 2009 B2
7556108 Won Jul 2009 B2
7556219 Page Jul 2009 B2
7587756 Peart Sep 2009 B2
7613544 Park Nov 2009 B2
7627515 Borgs Dec 2009 B2
7636045 Sugiyama Dec 2009 B2
7648068 Silverbrook Jan 2010 B2
7653603 Holtkamp, Jr. Jan 2010 B1
7658327 Tuchman Feb 2010 B2
7689322 Tanaka Mar 2010 B2
7693605 Park Apr 2010 B2
7693745 Pomerantz Apr 2010 B1
7693757 Zimmerman Apr 2010 B2
7706917 Chiappetta Apr 2010 B1
7716064 McIntyre May 2010 B2
7726563 Scott Jun 2010 B2
7762458 Stawar Jul 2010 B2
7783527 Bonner Aug 2010 B2
7787985 Tsujimoto Aug 2010 B2
7817394 Mukherjee Oct 2010 B2
7826919 DAndrea Nov 2010 B2
7835281 Lee Nov 2010 B2
7894932 Mountz Feb 2011 B2
7894939 Zini Feb 2011 B2
7969297 Haartsen Jun 2011 B2
7996109 Zini Aug 2011 B2
8010230 Zini Aug 2011 B2
8032249 Shakes Oct 2011 B1
8041455 Thorne Oct 2011 B2
8050976 Staib Nov 2011 B2
8065032 Stifter Nov 2011 B2
8065353 Eckhoff-Hornback Nov 2011 B2
8069092 Bryant Nov 2011 B2
8083013 Bewley Dec 2011 B2
8099191 Blanc Jan 2012 B2
8103398 Duggan Jan 2012 B2
8195333 Ziegler Jun 2012 B2
8239276 Lin Aug 2012 B2
8244041 Silver Aug 2012 B1
8248467 Ganick Aug 2012 B1
8260456 Siegel Sep 2012 B2
8284240 Saint-Pierre Oct 2012 B2
8295542 Albertson Oct 2012 B2
8321303 Krishnamurthy Nov 2012 B1
8325036 Fuhr Dec 2012 B1
8342467 Stachowski Jan 2013 B2
8352110 Szybalski Jan 2013 B1
8359122 Koselka Jan 2013 B2
8380349 Hickman Feb 2013 B1
8393846 Coots Mar 2013 B1
8412400 DAndrea Apr 2013 B2
8423280 Edwards Apr 2013 B2
8425173 Lert Apr 2013 B2
8429004 Hamilton Apr 2013 B2
8430192 Gillett Apr 2013 B2
8433470 Szybalski Apr 2013 B1
8433507 Hannah Apr 2013 B2
8437875 Hernandez May 2013 B2
8444369 Watt May 2013 B2
8447863 Francis, Jr. May 2013 B1
8452450 Dooley May 2013 B2
8474090 Jones Jul 2013 B2
8494908 Herwig Jul 2013 B2
8504202 Ichinose Aug 2013 B2
8508590 Laws Aug 2013 B2
8510033 Park Aug 2013 B2
8511606 Lutke Aug 2013 B1
8515580 Taylor Aug 2013 B2
8516651 Jones Aug 2013 B2
8538577 Bell Sep 2013 B2
8544858 Eberlein Oct 2013 B2
8571700 Keller Oct 2013 B2
8572712 Rice Oct 2013 B2
8577538 Lenser Nov 2013 B2
8587662 Moll Nov 2013 B1
8588969 Frazier Nov 2013 B2
8594834 Clark Nov 2013 B1
8606314 Barnes, Jr. Dec 2013 B2
8606392 Wurman Dec 2013 B2
8639382 Clark Jan 2014 B1
8645223 Ouimet Feb 2014 B2
8649557 Hyung Feb 2014 B2
8656550 Jones Feb 2014 B2
8670866 Ziegler Mar 2014 B2
8671507 Jones Mar 2014 B2
8676377 Siegel Mar 2014 B2
8676420 Kume Mar 2014 B2
8676480 Lynch Mar 2014 B2
8700230 Hannah Apr 2014 B1
8708285 Carreiro Apr 2014 B1
8718814 Clark May 2014 B1
8724282 Hiremath May 2014 B2
8732039 Chen May 2014 B1
8744626 Johnson Jun 2014 B2
8751042 Lee Jun 2014 B2
8763199 Jones Jul 2014 B2
8770976 Moser Jul 2014 B2
8775064 Zeng Jul 2014 B2
8798786 Wurman Aug 2014 B2
8798840 Fong Aug 2014 B2
8814039 Bishop Aug 2014 B2
8818556 Sanchez Aug 2014 B2
8820633 Bishop Sep 2014 B2
8825226 Worley, III Sep 2014 B1
8831984 Hoffman Sep 2014 B2
8838268 Friedman Sep 2014 B2
8843244 Phillips Sep 2014 B2
8851369 Bishop Oct 2014 B2
8882432 Bastian, II Nov 2014 B2
8886390 Wolfe Nov 2014 B2
8892240 Vliet Nov 2014 B1
8892241 Weiss Nov 2014 B2
8899903 Saad Dec 2014 B1
8918202 Kawano Dec 2014 B2
8918230 Chen Dec 2014 B2
8930044 Peeters Jan 2015 B1
8965561 Jacobus Feb 2015 B2
8972045 Mountz Mar 2015 B1
8972061 Rosenstein Mar 2015 B2
8983647 Dwarakanath Mar 2015 B1
8989053 Skaaksrud Mar 2015 B1
9002506 Agarwal Apr 2015 B1
9008827 Dwarakanath Apr 2015 B1
9008829 Worsley Apr 2015 B2
9014848 Farlow Apr 2015 B2
9075136 Joao Jul 2015 B1
9129277 MacIntosh Sep 2015 B2
9170117 Abuelsaad Oct 2015 B1
9173816 Reinhardt Nov 2015 B2
9190304 MacKnight Nov 2015 B2
1506144 High Mar 2016 A1
9278839 Gilbride Mar 2016 B2
9305280 Berg Apr 2016 B1
9329597 Stoschek May 2016 B2
1527500 High Sep 2016 A1
1527501 High Sep 2016 A1
1527504 High Sep 2016 A1
1528295 High Sep 2016 A1
1528892 High Oct 2016 A1
9495703 Kaye, III Nov 2016 B1
9534906 High Jan 2017 B2
9550577 Beckman Jan 2017 B1
1542381 High Feb 2017 A1
9573684 Kimchi Feb 2017 B2
9578282 Sills Feb 2017 B1
1544691 High Mar 2017 A1
1544717 High Mar 2017 A1
1544720 High Mar 2017 A1
1547127 High Mar 2017 A1
9607285 Wellman Mar 2017 B1
9623923 Riedel Apr 2017 B2
9649766 Stubbs May 2017 B2
9656805 Evans May 2017 B1
9658622 Walton May 2017 B2
9663292 Brazeau May 2017 B1
9663293 Wurman May 2017 B2
9663295 Wurman May 2017 B1
9663296 Dingle May 2017 B1
1569222 High Aug 2017 A1
9747480 Mcallister Aug 2017 B2
9757002 Thompson Sep 2017 B2
9801517 High Oct 2017 B2
9827678 Gilbertson Nov 2017 B1
1583670 High Dec 2017 A1
9875502 Kay Jan 2018 B2
9875503 High Jan 2018 B2
1589225 High Feb 2018 A1
1589415 High Feb 2018 A1
9896315 High Mar 2018 B2
9908760 High Mar 2018 B2
9994434 High Jun 2018 B2
20010042024 Rogers Nov 2001 A1
20020060542 Song May 2002 A1
20020095342 Feldman Jul 2002 A1
20020154974 Fukuda Oct 2002 A1
20020156551 Tackett Oct 2002 A1
20020165638 Bancroft Nov 2002 A1
20020165643 Bancroft Nov 2002 A1
20020165790 Bancroft Nov 2002 A1
20020174021 Chu Nov 2002 A1
20030028284 Chirnomas Feb 2003 A1
20030152679 Garwood Aug 2003 A1
20030170357 Garwood Sep 2003 A1
20030185948 Garwood Oct 2003 A1
20030222798 Floros Dec 2003 A1
20040068348 Jager Apr 2004 A1
20040081729 Garwood Apr 2004 A1
20040093650 Martins May 2004 A1
20040098167 Yi May 2004 A1
20040117063 Sabe Jun 2004 A1
20040146602 Garwood Jul 2004 A1
20040216339 Garberg Nov 2004 A1
20040217166 Myers Nov 2004 A1
20040221790 Sinclair Nov 2004 A1
20040225613 Narayanaswami Nov 2004 A1
20040249497 Saigh Dec 2004 A1
20050008463 Stehr Jan 2005 A1
20050047895 Lert Mar 2005 A1
20050072651 Wieth Apr 2005 A1
20050080520 Kline Apr 2005 A1
20050104547 Wang May 2005 A1
20050149414 Schrodt Jul 2005 A1
20050177446 Hoblit Aug 2005 A1
20050216126 Koselka Sep 2005 A1
20050222712 Orita Oct 2005 A1
20050230472 Chang Oct 2005 A1
20050238465 Razumov Oct 2005 A1
20060107067 Safal May 2006 A1
20060147087 Goncalves Jul 2006 A1
20060163350 Melton Jul 2006 A1
20060178777 Park Aug 2006 A1
20060206235 Shakes Sep 2006 A1
20060220809 Stigall Oct 2006 A1
20060221072 Se Oct 2006 A1
20060231301 Rose Oct 2006 A1
20060235570 Jung Oct 2006 A1
20060241827 Fukuchi Oct 2006 A1
20060244588 Hannah Nov 2006 A1
20060279421 French Dec 2006 A1
20060293810 Nakamoto Dec 2006 A1
20070005179 Mccrackin Jan 2007 A1
20070017855 Pippin Jan 2007 A1
20070045018 Carter Mar 2007 A1
20070061210 Chen Mar 2007 A1
20070085682 Murofushi Apr 2007 A1
20070125727 Winkler Jun 2007 A1
20070150368 Arora Jun 2007 A1
20070152057 Cato Jul 2007 A1
20070222679 Morris Sep 2007 A1
20070269299 Ross Nov 2007 A1
20070284442 Herskovitz Dec 2007 A1
20070288123 D Andrea Dec 2007 A1
20070293978 Wurman Dec 2007 A1
20080011836 Adema Jan 2008 A1
20080031491 Ma Feb 2008 A1
20080041644 Tudek Feb 2008 A1
20080042836 Christopher Feb 2008 A1
20080075566 Benedict Mar 2008 A1
20080075568 Benedict Mar 2008 A1
20080075569 Benedict Mar 2008 A1
20080077511 Zimmerman Mar 2008 A1
20080105445 Dayton May 2008 A1
20080131255 Hessler Jun 2008 A1
20080140253 Brown Jun 2008 A1
20080154720 Gounares Jun 2008 A1
20080201227 Bakewell Aug 2008 A1
20080226129 Kundu Sep 2008 A1
20080267759 Morency Oct 2008 A1
20080281515 Ann Nov 2008 A1
20080281664 Campbell Nov 2008 A1
20080294288 Yamauchi Nov 2008 A1
20080306787 Hamilton Dec 2008 A1
20080308630 Bhogal Dec 2008 A1
20080314667 Hannah Dec 2008 A1
20090074545 Lert Mar 2009 A1
20090132250 Chiang May 2009 A1
20090134572 Obuchi May 2009 A1
20090138375 Schwartz May 2009 A1
20090154708 Kolar Sunder Jun 2009 A1
20090155033 Olsen Jun 2009 A1
20090164902 Cohen Jun 2009 A1
20090210536 Allen Aug 2009 A1
20090240571 Bonner Sep 2009 A1
20090259571 Ebling Oct 2009 A1
20090265193 Collins Oct 2009 A1
20090269173 De Leo Oct 2009 A1
20090299822 Harari Dec 2009 A1
20090319399 Resta Dec 2009 A1
20100025964 Fisk Feb 2010 A1
20100030417 Fang Feb 2010 A1
20100076959 Ramani Mar 2010 A1
20100138281 Zhang Jun 2010 A1
20100143089 Hvass Jun 2010 A1
20100171826 Hamilton Jul 2010 A1
20100176922 Schwab Jul 2010 A1
20100211441 Sprigg Aug 2010 A1
20100222925 Anezaki Sep 2010 A1
20100268697 Karlsson Oct 2010 A1
20100295847 Titus Nov 2010 A1
20100299065 Mays Nov 2010 A1
20100302102 Desai Dec 2010 A1
20100324773 Choi Dec 2010 A1
20110010023 Kunzig Jan 2011 A1
20110022201 Reumerman Jan 2011 A1
20110098920 Chuang Apr 2011 A1
20110153081 Romanov Jun 2011 A1
20110163160 Zini Jul 2011 A1
20110176803 Song Jul 2011 A1
20110225071 Sano Sep 2011 A1
20110240777 Johns Oct 2011 A1
20110258060 Sweeney Oct 2011 A1
20110260865 Bergman Oct 2011 A1
20110279252 Carter Nov 2011 A1
20110288684 Farlow Nov 2011 A1
20110288763 Hui Nov 2011 A1
20110295424 Johnson Dec 2011 A1
20110301757 Jones Dec 2011 A1
20110320034 Dearlove Dec 2011 A1
20110320322 Roslak Dec 2011 A1
20120000024 Layton Jan 2012 A1
20120029697 Ota Feb 2012 A1
20120035823 Carter Feb 2012 A1
20120046998 Staib Feb 2012 A1
20120059743 Rao Mar 2012 A1
20120072303 Brown Mar 2012 A1
20120134771 Larson May 2012 A1
20120143726 Chirnomas Jun 2012 A1
20120192260 Kontsevich Jul 2012 A1
20120226556 Itagaki Sep 2012 A1
20120239224 McCabe Sep 2012 A1
20120255810 Yang Oct 2012 A1
20120259732 Sasankan Oct 2012 A1
20120272500 Reuteler Nov 2012 A1
20120294698 Villamar Nov 2012 A1
20120303263 Alam Nov 2012 A1
20120303479 Derks Nov 2012 A1
20120330458 Weiss Dec 2012 A1
20130016011 Harriman Jan 2013 A1
20130026224 Ganick Jan 2013 A1
20130051667 Deng Feb 2013 A1
20130054052 Waltz Feb 2013 A1
20130054280 Moshfeghi Feb 2013 A1
20130060461 Wong Mar 2013 A1
20130073405 Ariyibi Mar 2013 A1
20130096735 Byford Apr 2013 A1
20130103539 Abraham Apr 2013 A1
20130105036 Smith May 2013 A1
20130110671 Gray May 2013 A1
20130141555 Ganick Jun 2013 A1
20130145572 Schregardus Jun 2013 A1
20130151335 Avadhanam Jun 2013 A1
20130155058 Golparvar-Fard Jun 2013 A1
20130174371 Jones Jul 2013 A1
20130181370 Rafie Jul 2013 A1
20130211953 Abraham Aug 2013 A1
20130218453 Geelen Aug 2013 A1
20130235206 Smith Sep 2013 A1
20130238130 Dorschel Sep 2013 A1
20130276004 Boncyk Oct 2013 A1
20130300729 Grimaud Nov 2013 A1
20130302132 DAndrea Nov 2013 A1
20130309637 Minvielle Nov 2013 A1
20130317642 Asaria Nov 2013 A1
20130333961 ODonnell Dec 2013 A1
20130338825 Cantor Dec 2013 A1
20140006229 Birch Jan 2014 A1
20140014470 Razumov Jan 2014 A1
20140032034 Raptopoulos Jan 2014 A1
20140032379 Schuetz Jan 2014 A1
20140037404 Hancock Feb 2014 A1
20140046512 Villamar Feb 2014 A1
20140058556 Kawano Feb 2014 A1
20140067564 Yuan Mar 2014 A1
20140081445 Villamar Mar 2014 A1
20140091013 Streufert Apr 2014 A1
20140100715 Mountz Apr 2014 A1
20140100768 Kessens Apr 2014 A1
20140100769 Wurman Apr 2014 A1
20140100998 Mountz Apr 2014 A1
20140100999 Mountz Apr 2014 A1
20140101690 Boncyk Apr 2014 A1
20140108087 Fukui Apr 2014 A1
20140124004 Rosenstein May 2014 A1
20140129054 Huntzicker May 2014 A1
20140133943 Razumov May 2014 A1
20140135984 Hirata May 2014 A1
20140143039 Branton May 2014 A1
20140149958 Samadi May 2014 A1
20140152507 McAllister Jun 2014 A1
20140156450 Ruckart Jun 2014 A1
20140156461 Lerner Jun 2014 A1
20140157156 Kawamoto Jun 2014 A1
20140164123 Wissner-Gross Jun 2014 A1
20140172197 Ganz Jun 2014 A1
20140172727 Abhyanker Jun 2014 A1
20140177907 Argue Jun 2014 A1
20140177924 Argue Jun 2014 A1
20140180478 Letsky Jun 2014 A1
20140180528 Argue Jun 2014 A1
20140180865 Argue Jun 2014 A1
20140180914 Abhyanker Jun 2014 A1
20140201041 Meyer Jul 2014 A1
20140207614 Ramaswamy Jul 2014 A1
20140209514 Gitschel Jul 2014 A1
20140211988 Fan Jul 2014 A1
20140214205 Kwon Jul 2014 A1
20140217242 Muren Aug 2014 A1
20140228999 D'Andrea Aug 2014 A1
20140229320 Mohammed Aug 2014 A1
20140244026 Neiser Aug 2014 A1
20140244207 Hicks Aug 2014 A1
20140246257 Jacobsen Sep 2014 A1
20140247116 Davidson Sep 2014 A1
20140250613 Jones Sep 2014 A1
20140254896 Zhou Sep 2014 A1
20140257928 Chen Sep 2014 A1
20140266616 Jones Sep 2014 A1
20140274309 Nguyen Sep 2014 A1
20140277693 Naylor Sep 2014 A1
20140277742 Wells Sep 2014 A1
20140277841 Klicpera Sep 2014 A1
20140285134 Kim Sep 2014 A1
20140289009 Campbell Sep 2014 A1
20140297090 Ichinose Oct 2014 A1
20140304107 McAllister Oct 2014 A1
20140306654 Partovi Oct 2014 A1
20140309809 Dixon Oct 2014 A1
20140330456 LopezMorales Nov 2014 A1
20140330677 Boncyk Nov 2014 A1
20140344118 Parpia Nov 2014 A1
20140350725 LaFary Nov 2014 A1
20140350851 Carter Nov 2014 A1
20140350855 Vishnuvajhala Nov 2014 A1
20140361077 Davidson Dec 2014 A1
20140369558 Holz Dec 2014 A1
20140371912 Passot Dec 2014 A1
20140379588 Gates Dec 2014 A1
20150006319 Thomas Jan 2015 A1
20150029339 Kobres Jan 2015 A1
20150032252 Galluzzo Jan 2015 A1
20150045992 Ashby Feb 2015 A1
20150046299 Yan Feb 2015 A1
20150066283 Wurman Mar 2015 A1
20150073589 Khodl Mar 2015 A1
20150098775 Razumov Apr 2015 A1
20150100439 Lu Apr 2015 A1
20150100461 Baryakar Apr 2015 A1
20150112826 Crutchfield Apr 2015 A1
20150120094 Kimchi Apr 2015 A1
20150123973 Larsen May 2015 A1
20150142249 Ooga May 2015 A1
20150203140 Holtan Jul 2015 A1
20150205298 Stoschek Jul 2015 A1
20150205300 Caver Jul 2015 A1
20150217449 Meier Aug 2015 A1
20150217790 Golden Aug 2015 A1
20150221854 Melz Aug 2015 A1
20150229906 Inacio De Matos Aug 2015 A1
20150231873 Okamoto Aug 2015 A1
20150277440 Kimchi Oct 2015 A1
20150278889 Qian Oct 2015 A1
20150325128 Lord Nov 2015 A1
20150336668 Pasko Nov 2015 A1
20150360865 Massey Dec 2015 A1
20160016731 Razumov Jan 2016 A1
20160023675 Hannah Jan 2016 A1
20160052139 Hyde Feb 2016 A1
20160101794 Fowler Apr 2016 A1
20160101936 Chamberlin Apr 2016 A1
20160110701 Herring Apr 2016 A1
20160114488 Mascorro Medina Apr 2016 A1
20160167557 Mecklinger Jun 2016 A1
20160167577 Simmons Jun 2016 A1
20160176638 Toebes Jun 2016 A1
20160196755 Navot Jul 2016 A1
20160207193 Wise Jul 2016 A1
20160210602 Siddique Jul 2016 A1
20160236867 Brazeau Aug 2016 A1
20160255969 High Sep 2016 A1
20160257212 Thompson Sep 2016 A1
20160257240 High Sep 2016 A1
20160257401 Buchmueller Sep 2016 A1
20160258762 Taylor Sep 2016 A1
20160258763 High Sep 2016 A1
20160259028 High Sep 2016 A1
20160259329 High Sep 2016 A1
20160259331 Thompson Sep 2016 A1
20160259339 High Sep 2016 A1
20160259340 Kay Sep 2016 A1
20160259341 High Sep 2016 A1
20160259342 High Sep 2016 A1
20160259343 High Sep 2016 A1
20160259344 High Sep 2016 A1
20160259345 McHale Sep 2016 A1
20160259346 High Sep 2016 A1
20160260049 High Sep 2016 A1
20160260054 High Sep 2016 A1
20160260055 High Sep 2016 A1
20160260142 Winkle Sep 2016 A1
20160260145 High Sep 2016 A1
20160260148 High Sep 2016 A1
20160260158 High Sep 2016 A1
20160260159 Atchley Sep 2016 A1
20160260161 Atchley Sep 2016 A1
20160261698 Thompson Sep 2016 A1
20160274586 Stubbs Sep 2016 A1
20160288601 Gehrke Oct 2016 A1
20160300291 Carmeli Oct 2016 A1
20160301698 Katara Oct 2016 A1
20160325932 Hognaland Nov 2016 A1
20160349754 Mohr Dec 2016 A1
20160355337 Lert Dec 2016 A1
20160364785 Wankhede Dec 2016 A1
20160364786 Wankhede Dec 2016 A1
20170009417 High Jan 2017 A1
20170010608 High Jan 2017 A1
20170010609 High Jan 2017 A1
20170010610 Atchley Jan 2017 A1
20170020354 High Jan 2017 A1
20170024806 High Jan 2017 A1
20170080846 Lord Mar 2017 A1
20170107055 Magens Apr 2017 A1
20170110017 Kimchi Apr 2017 A1
20170120443 Kang May 2017 A1
20170129602 Alduaiji May 2017 A1
20170137235 Thompson May 2017 A1
20170148075 High May 2017 A1
20170158430 Raizer Jun 2017 A1
20170166399 Stubbs Jun 2017 A1
20170176986 High Jun 2017 A1
20170178066 High Jun 2017 A1
20170178082 High Jun 2017 A1
20170183159 Weiss Jun 2017 A1
20170283171 High Oct 2017 A1
20170355081 Fisher Dec 2017 A1
20180020896 High Jan 2018 A1
20180068357 High Mar 2018 A1
20180099846 High Apr 2018 A1
Foreign Referenced Citations (93)
Number Date Country
2524037 May 2006 CA
2625885 Apr 2007 CA
100999277 Jul 2007 CN
102079433 Jun 2011 CN
202847767 Apr 2013 CN
103136923 May 2013 CN
103213115 Jul 2013 CN
203166399 Aug 2013 CN
203191819 Sep 2013 CN
203401274 Jan 2014 CN
203402565 Jan 2014 CN
103625808 Mar 2014 CN
203468521 Mar 2014 CN
103696393 Apr 2014 CN
103723403 Apr 2014 CN
203512491 Apr 2014 CN
103770117 May 2014 CN
203782622 Aug 2014 CN
104102188 Oct 2014 CN
104102219 Oct 2014 CN
102393739 Dec 2014 CN
204054062 Dec 2014 CN
204309852 Dec 2014 CN
204331404 May 2015 CN
105460051 Apr 2016 CN
102013013438 Feb 2015 DE
861415 May 1997 EP
1136052 Sep 2001 EP
0887491 Apr 2004 EP
1439039 Jul 2004 EP
1447726 Aug 2004 EP
2148169 Jan 2010 EP
2106886 Mar 2011 EP
2309487 Apr 2011 EP
2050544 Aug 2011 EP
2498158 Sep 2012 EP
2571660 Mar 2013 EP
2590041 May 2013 EP
2608163 Jun 2013 EP
2662831 Nov 2013 EP
2730377 May 2014 EP
2710330 Mar 1995 FR
1382806 Feb 1971 GB
2530626 Mar 2016 GB
2542472 Mar 2017 GB
2542905 May 2017 GB
62247458 Oct 1987 JP
H10129996 May 1998 JP
2003288396 Oct 2003 JP
2005350222 Dec 2005 JP
2009284944 Dec 2009 JP
2010105644 May 2010 JP
2010231470 Oct 2010 JP
20120100505 Sep 2012 KR
8503277 Aug 1985 WO
9603305 Jul 1995 WO
1997018523 May 1997 WO
9855903 Dec 1998 WO
2000061438 Oct 2000 WO
0132366 May 2001 WO
2004092858 Oct 2004 WO
2005102875 Nov 2005 WO
2006056614 Jun 2006 WO
2006120636 Nov 2006 WO
2006137072 Dec 2006 WO
2007007354 Jan 2007 WO
2007047514 Apr 2007 WO
2007149196 Dec 2007 WO
2008118906 Oct 2008 WO
2008144638 Nov 2008 WO
2008151345 Dec 2008 WO
2009022859 Feb 2009 WO
2009027835 Mar 2009 WO
2009103008 Aug 2009 WO
2011063527 Jun 2011 WO
2012075196 Jun 2012 WO
2013138193 Sep 2013 WO
2013138333 Sep 2013 WO
2013176762 Nov 2013 WO
2014022366 Feb 2014 WO
2014022496 Feb 2014 WO
2014045225 Mar 2014 WO
2014046757 Mar 2014 WO
2014101714 Jul 2014 WO
2014116947 Jul 2014 WO
2014138472 Sep 2014 WO
2014165286 Oct 2014 WO
2015021958 Feb 2015 WO
2015104263 Jul 2015 WO
2015155556 Oct 2015 WO
2016009423 Jan 2016 WO
2016015000 Jan 2016 WO
2016144765 Sep 2016 WO
Non-Patent Literature Citations (154)
Entry
“SK Telecom Launches Smart Cart Pilot Test in Korea” online article at http://www.sktelecom.com dated Oct. 4, 2011.
“Hop the hands-free suitcase follows you around like an obedient pet” online article at https://www.digitaltrends.com/cool-tech/hop-the-hands-free-suitcase-follows-you-around-l . . . published in 2011.
“A Control System for the Robot Shopping Cart” by Onozato et al presented in 2010 IRAST International Congress on Computer Applications and Computational Science (CACS 2010).
Bohren; Jonathan et al.; “Towards Autonomous Robotic Butlers: Lessons Learned with the PR2”, Willow Garage, May 9, 2011, pp. 1-8.
Denso; “X-mobility”, Oct. 10, 2014, pp. 1-2, including machine translation.
Fellow Robots; “Meet Oshbot”, http://fellowrobots.com/oshbot/, May 19, 2015, pp. 1-3.
Green, A., et al; “Report on evaluation of the robot trolley”, CommRob IST-045441, Advanced Behaviour and High-Level Multimodal Communications with and among Robots, Jun. 14, 2010, pp. 10-67.
Koubaa, Anis; “A Service-Oriented Architecture for Virtualizing Robots in Robot-as-a-Service Clouds”, 2014, pp. 1-13.
Kumar Paradkar, Prashant; “Voice Controlled Robotic Project using interfacing of Ardruino and Bluetooth HC-05”, Robotics_Projects_C/C++_Android, Jan. 23, 2016, pp. 1-14.
Meena, M., & Thilagavathi, P.; “Automatic Docking System with Recharging and Battery Replacement for Surveillance Robot”, International Journal of Electronics and Computer Science Engineering, 2012, pp. 1148-1154.
Sebaali, G., et al.; “Smart Shopping Cart”, Department of Electrical and Computer Engineering, American University of Beirut, 2014, pp. 1-6.
Shukla, Neha; “SaviOne the Butler Bot: Service Robot for Hospitality Industry”, TechieTonics, http://www.techietonics.com/robo-tonics/savione-the-butler-bot-service-for-hospitality-industry.html, Aug. 14, 2014, pp. 1-5.
Swisslog; “RoboCourier Autonomous Mobile Robot”, http://www.swisslog.com/en/Products/HCS/Automated-Material-Transport/RoboCourier-Autonomous-Mobile-Robot, printed May 27, 2015, pp. 1.
Vasilescu, Iuliu, et al.; “Autonomous Modular Optical Underwater Robot (AMOUR) Design, Prototype and Feasibility Study”, Apr. 18, 2005, pp. 1-7.
Wikipedia; “Leeds Kirkgate Market”; https://en.wikipedia.org/wiki/Leeds_Kirkgate_Market; Retrieved on Apr. 5, 2017; 8 pages.
U.S. Appl. No. 15/060,953, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,025, filed Mar. 4, 2016, Kay.
U.S. Appl. No. 15/061,054, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,203, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,265, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,285, filed Mar. 4, 2016, Thompson.
U.S. Appl. No. 15/061,325, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,350, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,402, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,406, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,474, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,507, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,671, filed Mar. 4, 2016, Taylor.
U.S. Appl. No. 15/061,677, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,686, filed Mar. 4, 2016, Thompson.
U.S. Appl. No. 15/061,688, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,722, filed Mar. 4, 2016, Atchley.
U.S. Appl. No. 15/061,770, filed Mar. 4, 2016, Winkle.
U.S. Appl. No. 15/061,792, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,801, filed Mar. 4, 2016, Atchley.
U.S. Appl. No. 15/061,805, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,844, filed Mar. 4, 2016, High.
U.S. Appl. No. 15/061,848, filed Mar. 4, 2016, McHale.
U.S. Appl. No. 15/061,908, filed Mar. 4, 2016, High.
ABBRobotics; “ABB Robotics—Innovative Packaging Solutions”, https://www.youtube.com/watch?v=e5jif-IUvHY, published on May 16, 2013, pp. 1-5.
Ang, Fitzwatler, et al.; “Automated Waste Sorter With Mobile Robot Delivery Waste System”, De La Salle University Research Congress 2013, Mar. 7-9, 2013, pp. 1-7.
Ansari, Sameer, et al.; “Automated Trash Collection & Removal in Office Cubicle Environments”, Squad Collaborative Robots, Sep. 27, 2013, pp. 1-23.
Armstrong, Jean, et al.; “Visible Light Positioning: A Roadmap for International Standardization”, IEEE Communications Magazine, Dec. 2013, pp. 2-7.
Artal, J.S., et al.; “Autonomous Mobile Robot with Hybrid PEM Fuel-Cell and Ultracapacitors Energy System, Dedalo 2.0”, International Conference on Renewable Energies and Power Quality, Santiago de Compostela, Spain, Mar. 28-30, 2012, pp. 1-6.
Atherton, Kelsey D.; “New GPS Receiver Offers Navigation Accurate to an Inch”, Popular Science, www.popsci.com/technology/article/2013-08/global-positioning-down-inches, Aug. 16, 2013, pp. 1-2.
Avezbadalov, Ariel, et al.; “Snow Shoveling Robot”, engineering.nyu.edu/mechatronics/projects/ME3484/2006/Snow Shoveling Robot/Mechatronics Snow Robot Presentation Update 12-19-06.pdf, 2006, pp. 1-24.
Bares, John, et al.; “Designing Crash-Survivable Unmanned Vehicles”, AUVSI Symposium, Jul. 10, 2002, pp. 1-15.
Bohren; Jonathan et al.; “Towards Autonomous Robotic Butlers: Lessons Learned with the PR2”, Willow Garage, pp. 1-8.
Bouchard, Samuel; “A Robot to Clean Your Trash Bin!”, Robotiq, http://blog.robotiq.com/bid/41203/A-Robot-to-Clean-your-Trash-Bin, Aug. 22, 2011, pp. 1-7.
Burns, Tom; “irobot roomba 780 review best robot vacuum floor cleaning robot review video demo”, https://www.youtube.com/watch?v=MkwtlyVAaEY, published on Feb. 13, 2013, pp. 1-10.
ByteLight; “Scalable Indoor Location”, http://www.bytelight.com/, Dec. 12, 2014, pp. 1-2.
Canadian Manufacturing; “Amazon unleashes army of order-picking robots”, http://www.canadianmanufacturing.com/supply-chain/amazon-unleashes-army-order-picking-robots-142902/, Dec. 2, 2014, pp. 1-4.
Capel, Claudine; “Waste sorting—A look at the separation and sorting techniques in today's European market”, Waste Management World, http://waste-management-world.com/a/waste-sorting-a-look-at-the-separation-and-sorting-techniques-in-todayrsquos-european-market, Jul. 1, 2008, pp. 1-8.
Carnegie Mellon Univeristy; “AndyVision—The Future of Retail”, https://www.youtube.com/watch?v=n5309ILTV2s, published on Jul. 16, 2012, pp. 1-9.
Carnegie Mellon University; “Robots in Retail”, www.cmu.edu/homepage/computing/2012/summer/robots-in-retail.shmtl, 2012, pp. 1.
Chopade, Jayesh, et al.; “Control of Spy Robot by Voice and Computer Commands”, International Journal of Advanced Research in Computer and Communication Engineering, vol. 2, Issue 4, Apr. 2013, pp. 1-3.
CNET; “iRobot Braava 380t—No standing ovation for this robotic floor mop”, https://www.youtube.com/watch?v=JAtClxFtC6Q, published on May 7, 2014, pp. 1-6.
Coltin, Brian & Ventura, Rodrigo; “Dynamic User Task Scheduling for Mobile Robots”, Association for the Advancement of Artificial Intelligence, 2011, pp. 1-6.
Couceiro, Micael S., et al.; “Marsupial teams of robots: deployment of miniature robots for swarm exploration under communication constraints”, Robotica, Cambridge University Press, downloaded Jan. 14, 2014, pp. 1-22.
Coxworth, Ben; “Robot designed to sort trash for recycling”, Gizmag, http://www.gizmag.com/robot-sorts-trash-for-recycling/18426/, Apr. 18, 2011, pp. 1-7.
Davis, Jo; “The Future of Retail: In Store Now”, Online Brands, http://onlinebrands.co.nz/587/future-retail-store-now/, Nov. 16, 2014, pp. 1-5.
Denso; “X-mobility”, pp. 1.
DHL; “Self-Driving Vehicles in Logistics: A DHL perspective on implications and use cases for the logistics industry”, 2014, pp. 1-39.
Dorrier, Jason; “Service Robots Will Now Assist Customers at Lowe's Store”, SingularityHUB, http://singularityhub.com/2014/10/29/service-robots-will-now-assist-customers-at-lowes-store/, Oct. 29, 2014, pp. 1-4.
DroneWatch; “Weatherproof Drone XAircraft Has ‘Black Box’”, DroneWatch, http://www.dronewatch.nl/2015/02/13/weatherproof-drone-van-xaircraft-beschikt-over-zwarte-doos/, Feb. 13, 2015, pp. 1-5.
Dyson US; “See the new Dyson 360 Eye robot vacuum cleaner in action #DysonRobot”, https://www.youtube.com/watch?v=OadhulCDAjk, published on Sep. 4, 2014, pp. 1-7.
Edwards, Lin; “Supermarket robot to help the elderly (w/Video)”, Phys.Org, http://phys.org/news/2009-12-supermarket-robot-elderly-video.html, Dec. 17, 2009, pp. 1-5.
Elfes, Alberto; “Using Occupancy Grids for Mobile Robot Perception and Navigation”, IEEE, 1989, pp. 46-57.
Elkins, Herschel T.; “Important 2014 New Consumer Laws”, County of Los Angeles Department of Consumer Affairs Community Outreach & Education, updated Jan. 6, 2014, pp. 1-46.
Falconer, Jason; “HOSPI-R drug delivery robot frees nurses to do more important work”, Gizmag, http://www.gizmag.com/panasonic-hospi-r-delivery-robot/29565/, Oct. 28, 2013, pp. 1-6.
Falconer, Jason; “Toyota unveils helpful Human Support Robot”, Gizmag, http:/www.gizmag.com/toyota-human-support-robot/24246/, Sep. 22, 2012, pp. 1-6.
Farivar, Cyrus; “This in-store robot can show you the hammer aisle, but not the bathroom”, Ars Technica, http://arstechnica.com/business/2014/12/this-in-store-robot-can-show-you-the-hammer-aisle-but-not-the-bathroom/, Dec. 3, 2014, pp. 1-4.
Fellow Robots; “Meet Oshbot”, http://fellowrobots.com/oshbot/, pp. 1-3.
FellowRobots; “Oshbot Progress—Fellow Robots”, https://vimeo.com/139532370, published Sep. 16, 2015, pp. 1-5.
Fora.tv; “A Day in the Life of a Kiva Robot”, https://www.youtube.com/watch?v=6KRjuuEVEZs, published on May 11, 2011, pp. 1-11.
Gamma2video; “FridayBeerBot.wmv”, https://www.youtube.com/watch?v=KXXIIDYatxQ, published on Apr. 27, 2010, pp. 1-7.
Glas, Dylan F., et al.; “The Network Robot System: Enabling Social Human-Robot Interaction in Public Spaces”, Journal of Human-Robot Interaction, vol. 1, No. 2, 2012, pp. 5-32.
Green, A., et al; “Report on evaluation of the robot trolley”, CommRob IST-045441, Advanced Behaviour and High-Level Multimodal Communications with and among Robots, pp. 10-67.
Gross, H.-M., et al.; TOOMAS: Interactive Shopping Guide Robots in Everyday Use—Final Implementation and Experiences from Long-term Field Trials, Proc. IEEE/RJS Intern. Conf. on Intelligent Robots and Systems (IROS'09), St. Louis, USA, pp. 2005-2012.
Habib, Maki K., “Real Time Mapping and Dynamic Navigation for Mobile Robots”, International Journal of Advanced Robotic Systems, vol. 4, No. 3, 2007, pp. 323-338.
HRJ3 Productions; “Japanese Automatic Golf Cart”, https://www.youtube.com/watch?v=8diWYtqb6C0, published on Mar. 29, 2014, pp. 1-4.
Huang, Edward Y.C.; “A Semi-Autonomous Vision-Based Navigation System for a Mobile Robotic Vehicle”, Thesis submitted to the Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science on May 21, 2003, pp. 1-76.
IEEE Spectrum; “Warehouse Robots at Work”, https://www.youtube.com/watch?v=IWsMdN7HMuA, published on Jul. 21, 2008, pp. 1-11.
Intelligent Autonomous Systems; “TUM James goes shopping”, https://www.youtube.com/watch?v=JS2zycc4AUE, published on May 23, 2011, pp. 1-13.
Katic, M., Dusko; “Cooperative Multi Robot Systems for Contemporary Shopping Malls”, Robotics Laboratory, Mihailo Pupin Institute, University of Belgrade, Dec. 30, 2010, pp. 10-17.
Kehoe, Ben, et al.; “Cloud-Based Robot Grasping with the Google Object Recognition Engine”, 2013, pp. 1-7.
Kendricks, Cooper; “Trash Disposal Robot”, https://prezi.com31acae05zf8i/trash-disposal-robot/, Jan. 9, 2015, pp. 1-7.
Kibria, Shafkat, “Speech Recognition for Robotic Control”, Master's Thesis in Computing Science, Umea University, Dec. 18, 2005, pp. 1-77.
King, Rachael; “Newest Workers for Lowe's: Robots”, The Wall Street Journal, http:/www.wsj.com/articles/newest-workers-for-lowes-robots-1414468866, Oct. 28, 2014, pp. 1-4.
Kitamura, Shunichi; “Super Golf Cart with Remote drive and NAVI system in Japan”, https://www.youtube.com/watch?v=2_3-dUR12F8, published on Oct. 4, 2009, pp. 1-6.
Kiva Systems; “Automated Goods-to-Man Order Picking System—Kiva Systems”, http://www.kivasystems.com/solutions/picking/, printed on Apr. 2, 2015, pp. 1-2.
Kiva Systems; “Frequently Asked Questions about Kiva Systems—Kiva Systems”, http://kivasystems.com/about-us-the-kiva-approach/faq/, printed on Apr. 2, 2015, pp. 1-2.
Kiva Systems; “how a Kiva system makes use of the vertical space—Kiva Systems”, http://www.kivasystems.com/solutions/vertical-storage/, printed on Apr. 2, 2015, pp. 1-2.
Kiva Systems; “How Kiva Systems and Warehouse Management Systems Interact”, 2010, pp. 1-12.
Kiva Systems; “Kiva's warehouse automation system is the most powerful and flexible A . . . ”, http://www.kivasystems.com/solutions/, printed on Apr. 2, 2015, pp. 1-2.
Kiva Systems; “Kiva replenishment is more productive and accurate than replenishing pick faces in traditional distribution operations”, http//www.kivasystems.com/solutions/replenishment/, printed on Apr. 2, 2015, pp. 1-2.
Kiva Systems; “Kiva warehouse control software, Kiva WCS—Kiva Systems”, http://www.kivasystems.com/solutions/software/, printed on Apr. 2, 2015, pp. 1-2.
Kiva Systems; “Shipping Sortation—Kiva Systems”, http://www.kivasystems.com/solutions/shipping-sortation/, printed on Apr. 2, 2015, pp. 1-2.
Kohtsuka, Takafumi, et al.; “Design of a Control System for Robot Shopping Carts”, Knowledge-Based and Intelligent Information and Engineering Systems, 15th International Conference, KES 2011, Kaiserslautern, Germany, Sep. 12-14, 2011, pp. 280-288.
Koubaa, Anis; “A Service-Oriented Architecture for Virtualizing Robots in Robot-as-a-Service Clouds”, pp. 1-13.
Kumar, Swagat; “Robotics-as-a-Service: Transforming the Future of Retail”, Tata Consultancy Services, http://www.tcs.com/resources/white_papers/Pages/Robotics-as-Service.aspx, printed on May 13, 2015, pp. 1-4.
Kumar Paradkar, Prashant; “Voice Controlled Robotic Project using interfacing of Andruino and Bluetooth HC-05”, Robotics_Projects_C/C++_Android.
Lejepekov, Fedor; “Yuki-taro. Snow recycle robot.”, https://www.youtube.com/watch?v=gl2j9PY4jGY, published on Jan. 17, 2011, pp. 1-4.
Liu, Xiaohan, et al.; “Design of an Indoor Self-Positioning System for the Visually Impaired—Simulation with RFID and Bluetooth in a Visible Light Communication System”, Proceedings of the 29th Annual International Conference of the IEEE EMBS, Cite Internationale, Lyon, France, Aug. 23-26, 2007, pp. 1655-1658.
Lowe's Home Improvement; “OSHbots from Lowe's Innovation Labs”, https://www.youtube.com/watch?v=W-RKAjP1dtA, published on Dec. 15, 2014, pp. 1-8.
Lowe's Innovation Labs; “Autonomous Retail Service Robots”, http://www.lowesinnovationlabs.com/innovation-robots/, printed on Feb. 26, 2015, pp. 1-4.
Matos, Luis; “wi-GO—The autonomous and self-driven shopping cart”; https://www.indiegogo.com/projects/wi-go-the-autonomous-and-self-driven-shopping-cart; printed on Feb. 27, 2015, pp. 1-16.
Meena, M., & Thilagavahi, P.; “Automatic Docking System with Recharging and Battery Replacement for Surveillance Robot”, International Journal of Electronics and Computer Science Engineering, pp. 1148-1154.
Murph, Darren; “B.O.S.S. shopping cart follows you around”, Engadget, http://www.engadget.com/2006/08/11/b-o-s-s-shopping-cart-follows-you-around/, Aug. 11, 2006, pp. 1-4.
Nakajima, Madoka & Haruyama, Shinichiro; “New indoor navigation system for visually impaired people using visible light communication”, EURASIP Journal on Wireless Communications and Networking, 2013, pp. 1-10.
NEUROBTV; “Shopping Robot TOOMAS 2009”, https://www.youtube.com/watch?v=49Pkm30qmQU, published on May 8, 2010, pp. 1-7.
Nickerson, S.B., et al.; “An autonomous mobile robot for known industrial environments”, Autonomous Robot for a Known environment, Aug. 28, 1997, pp. 1-28.
O'Donnell, Jake; “Meet the Bluetooth-Connected Self-Following Robo-Caddy of the Future”, Sportsgrid; http://www.sportsgrid.com/uncategorized/meet-the-bluetooth-connected-self-following-robo-caddy-of-the-future/, Apr. 22, 2014, pp. 1-5.
Ogawa, Keisuke; “Denso Demos In-wheel Motor System for Baby Carriages, Shopping Carts”, Nikkei Technology, http://techon.nikkeiibp.co.jp/english/NEWS_EN/20141010/381880/?ST=english_PRINT, Oct. 10, 2014, pp. 1-2.
Orchard Supply Hardware; “Orchard Supply Hardware's OSHbot”, https://www.youtube.com/watch?v=Sp9176vm7Co, published on Oct. 28, 2014, pp. 1-9.
Osborne, Charlie; “Smart Cart Follows You When Grocery Shopping”, Smartplanet, http://www.smartplanet.com/blog/smart-takes/smart-cart-follows-you-when-grocery-shopping/, Feb. 29, 2012, pp. 1-4.
Poudel, Dev Bahadur; “Coordinating Hundreds of Cooperative, Autonomous Robots in a Warehouse”, Jan. 27, 2013, pp. 1-13.
Robotlab Inc.; “NAO robot drives autonomously it's own car”, https://www.youtube.com/watch?v=oBHYwYlo1UE, published on Sep. 8, 2014, pp. 1-6.
Rodriguez, Ashley; “Meet Lowe's Newest Sales Associate—OSHbot, the Robot”, Advertising Age, http://adage.com/article/cmo-strategy/meet-lowe-s-newest-sales-associate-oshbot-robot/295591/, Oct. 28, 2014, pp. 1-8.
Sebaali, G., et al.; “Smart Shopping Cart”, Department of Electrical and Computer Engineering, American University of Beirut, pp. 1-6.
Shukla, Neha; “SaviOne the Butler Bot: Service Robot for Hospitality Industry”, TechieTonics, http://www.techietonics.com/robo-tonics/savione-the-butler-bot-service-for-hospitality-industry.html, pp. 1-5.
Song, Guangming, et al.; “Automatic Docking System for Recharging Home Surveillance Robots”, http://www.academia.edu/6495007/Automatic_Docking_System_for_Recharging_Home_Surveillance_Robots, IEEE Transactions on Consumer Electronics, vol. 57, No. 2, May 2011, pp. 1-8.
Soper, Taylor; “Amazon vet's new robot-powered apparel startup aims to revolutionize how we buy clothes”, GeekWire, http://www.geekwire.com/2012/hointer-robot-jeans-clothing-apparel-store-startup/, Nov. 29, 2012, pp. 1-12.
Stewart Golf; “Introducing the New Stewart Golf X9 Follow”, https://www.youtube.com/watch?v=HHivFGtiuE, published on Apr. 9, 2014, pp. 1-9.
Sun, Eric; ““Smart Bin & Trash Route” system—RMIT 2012 Green Inventors Competition”, http://www.youtube.com/watch?v=OrTA57alO0k, published on Nov. 14, 2012, pp. 1-8.
Superdroid Robots; “Cool Robots, Making Life Easier”, http://www.superdroidrobots.com/shop/custom.aspx/cool-robots-making-life-easier/83/, printed on Jun. 16, 2015, pp. 1-7.
Swisslog; “RoboCourier Autonomous Mobile Robot”, http://www.swisslog.com/en/Products/HCS/Automated-Material-Transport/RoboCourier-Autonomous-Mobile-Robot, pp. 1.
Tam, Donna; “Meet Amazon's busiest employee—the Kiva robot”, CNET, http://www.cnet.com/news/meet-amazons-busiest-employee-the-kiva-robot/, Nov. 30, 2014, pp. 1-6.
Universal Robotics; “Neocortex Enables Random Part Handling and Automated Assembly”, http://www.universalrobotics.com/random-bin-picking, printed on Dec. 22, 2015, pp. 1-3.
UPHIGH Productions; “Behold the Future (E017 Robot Sales Assistant)”, https://www.youtube.com/watch?v=8WbvjaPm7d4, published on Nov. 19, 2014, pp. 1-7.
Urankar, Sandeep, et al.; “Robo-Sloth: A Rope-Climbing Robot”, Department of Mechanical Engineering, Indian Institute of Technology, 2003, pp. 1-10.
Vasilescu, Iuliu, et al.; “Autonomous Modular Optical Underwater Robot (AMOUR) Design, Prototype and Feasibility Study”, pp. 1-7.
VMECAVacuumTech; “VMECA Magic Suction Cup with ABB robot for pick and place (packaging application)”, https://www.youtube.com/watch?v=5btR9MLtGJA, published on Sep. 14, 2014, pp. 1-4.
Wang, Xuan; “2D Mapping Solutions for Low Cost Mobile Robot”, Master's Thesis in Computer Science, Royal Institute of Technology, KTH CSC, Stockholm, Sweden, 2013, pp. 1-60.
Webb, Mick; “Robovie II—the personal robotic shopping”, Gizmag, http://www.gizmag.com/robovie-ii-robotic-shopping-assistance/13664/, Dec. 23, 2009, pp. 1-5.
Weise, Elizabeth; “15,000 robots usher in Amazon's Cyber Monday”, USATODAY, http://www.usatoday.com/story/tech/2014/12/01/robots-amazon.kiva-fulfillment-centers-cyber-monday/19725229/, Dec. 2, 2014, pp. 1-3.
Weiss, C.C.; “Multifunctional hybrid robot shovels snow and mows your lawn”, Gizmag, http://www.gizmag.com/snowbyte-snow-shoveling-robot/32961/, Jul. 21, 2014, pp. 1-7.
Wikipedia; “Kiva Systems”, http://en.wikipedia.org/wiki/Kiva_Systems, printed on Apr. 2, 2015, pp. 1-3.
WIRED; “High-Speed Robots Part 1: Meet BettyBot in “Human Exclusion Zone” Warehouses—The Window—WIRED”, https://www.youtube.com/watch?v=8gy5tYVR-28, published on Jul. 2, 2013, pp. 1-6.
Wulf, O., et al.; “Colored 2D maps for robot navigation with 3D sensor data,” Institute for Systems Engineering, University of Hannover, Hannover, Germany, 2014, pp. 1-6.
YRF; “The Diamond Robbery—Scene Dhoom:2 Hrithik Roshan”, https://www.youtube.com/watch?v=3bMYgo_S0Kc, published on Jul. 12, 2012, pp. 1-7.
Daily Mail; “Dancing with your phone: The gyrating robotic dock that can move along with your music”, Sep. 12, 2012, http://www.dailymail.co.uk/sciencetech/article-2202164/The-intelligent-dancing-robot-controlled-mobile-phone.html, pp. 1-23.
Messieh, Nancy; “Humanoid robots will be roaming Abu Dhabi's malls next year”, The Next Web, Oct. 17, 2011, https://thenextweb.com/me/2011/10/17/humanoid-robots-will-be-roaming-abu-dhabis-malls-next-year/, pp. 1-6.
Owano, Nancy; “HEARBO robot can tell beeps, notes, and spoken word (w/ Video)”, Phys.org, Nov. 21, 2012, https://phys.org/news/2012-11-hearbo-robot-beeps-spoken-word.html, pp. 1-4.
Sales, Jorge, et al.; “CompaRob: The Shopping Cart Assistance Robot”, International Journal of Distributed Sensor Networks, vol. 2016, Article ID 4781280, Jan. 3, 2016, http://dx.doi.org/10.1155/2016/4781280, pp. 1-16.
Budgee; “The Robotic Shopping Cart Budgee ”; https://www.youtube.com/watch?v=2dYNdVPF4VM; published on Mar. 20, 2015; pp. 1-6.
Follow Inspiration; “wiiGO”; https://www.youtube.com/watch?v=dhHXIdpknC4; published on Jun. 16, 2015; pp. 1-7.
Technion; “Autonomous Tracking Shopping Cart—Shopping Made Easy from Technion”; https://www.youtube.com/watch?v=pQcb9fofmXg; published on Nov. 23, 2014; pp. 1-10.
Garun, Natt; “Hop the hands-free suitcase follows you around like an obedient pet”; https://www.digitaltrends.com/cool-tech/hop-the-hands-free-suitcase-follows-you-around-like-an-obedient-pet/; Oct. 10, 2012; pp. 1-6.
Kohtsuka, T. et al.; “Design of a Control System for Robot Shopping Carts”; KES'11 Proceedings of the 15th International Conference on Knowledge-Based and Intelligent Information and Engineering Systems; Sep. 12-14, 2011; pp. 280-288.
Nishimura, S. et al.; “Development of Attachable Modules for Robotizing Daily Items: Person Following Shopping Cart Robot”; Proceedings of the 2007 IEEE International Conference on Robotics and Biomimetics (Sanya, China); Dec. 15-18, 2007; pp. 1506-1511.
Onozato, Taishi et al.; “A Control System for the Robot Shopping Cart”; 2010 IRAST International Congress on Computer Applications and Computational Science (CACS 2010); 2010; pp. 907-910.
Scholz, J. et al.; “Cart Pushing with a Mobile Manipulation System: Towards Navigation with Moveable Objects”; Proceedings of the 2011 IEEE International Conference on Robotics and Automation (Shanghai, China); May 9-13, 2011; pp. 6115-6120.
SK Telecom Co.; “SK Telecom Launches Smart Cart Pilot Test in Korea”; http://www.sktelecom.com/en/press/press_detail.do?idx=971; Oct. 4, 2011; pp. 1-2.
Related Publications (1)
Number Date Country
20160261698 A1 Sep 2016 US
Provisional Applications (37)
Number Date Country
62129726 Mar 2015 US
62129727 Mar 2015 US
62138877 Mar 2015 US
62138885 Mar 2015 US
62152421 Apr 2015 US
62152465 Apr 2015 US
62152440 Apr 2015 US
62152630 Apr 2015 US
62152711 Apr 2015 US
62152610 Apr 2015 US
62152667 Apr 2015 US
62157388 May 2015 US
62165579 May 2015 US
62165416 May 2015 US
62165586 May 2015 US
62171822 Jun 2015 US
62175182 Jun 2015 US
62182339 Jun 2015 US
62185478 Jun 2015 US
62194131 Jul 2015 US
62194119 Jul 2015 US
62194121 Jul 2015 US
62194127 Jul 2015 US
62202744 Aug 2015 US
62202747 Aug 2015 US
62205548 Aug 2015 US
62205569 Aug 2015 US
62205555 Aug 2015 US
62205539 Aug 2015 US
62207858 Aug 2015 US
62214826 Sep 2015 US
62214824 Sep 2015 US
62292084 Feb 2016 US
62302547 Mar 2016 US
62302567 Mar 2016 US
62302713 Mar 2016 US
62303021 Mar 2016 US