The present disclosure pertains to robotic systems and methods that utilize artificial intelligence and, more particularly, to beverage and food preparation in an open environment using computer vision to locate and manipulate utensils, cookware, and ingredients that do not have a fixed location, and prepare and serve the beverage and food to the consumer with little to no human interaction. The present disclosure further relates to the field of using a visual guided system for a robot to perform food and beverage preparation and service tasks in a manner that is very similar to those of experienced human baristas.
Current methods for making specialty coffee drinks involve one or more human baristas to operate a commercial coffee machine. Typical steps of operation include grinding the coffee and collecting the ground coffee in a portable filter basket (known as a portafilter), tamping the coffee in the portafilter, inserting and locking the portafilter in the coffee machine, preparing the coffee by running hot water through the portafilter basket to extract a liquid “shot” of coffee, and finally, pouring the liquid coffee in a cup where it may be mixed with other ingredients. For a milk drink, additional steps include steaming and frothing the milk in a pitcher, and pouring the steamed milk into a cup. Optional latte art may be added to the very top surface of the drink. All of the foregoing steps or tasks are repetitive movements that require basic but consistent motor functions.
Attempts have been made to utilize robotics for performing the foregoing tasks. However, such attempts generally require the system to be in a closed box, and as such items and tools in the environment are fixed in location. This allows one to program robots to perform predefined and hardcoded motions that can complete these tasks given that the placement of the items do not change. However, in realistic working conditions in coffee shops, utensils and cookware are constantly moved about by baristas, and even with a robotic assistant, there is the issue of the robotic assistant being able to locate the correct utensil for the desired task.
Some other attempts have been made to utilize computer vision for performing some of tasks previously performed by humans. However, these attempts focus on recognizing and locating certain objects, instead of on how to modify robot motions to account for a constantly changing environment.
Attempts have been made to utilize computer vision for monitoring the environment and identifying conditions to act on. However, these attempts focus on raising alarms for critical conditions, instead of how to modify robot motions to resolve such situations.
Other attempts have been made to utilize computer vision for human recognition and understanding. However, these attempts focus on learning human behavior, instead of driving robots to interact with humans.
The present disclosure is directed to a modular robotic coffee barista station that is configured to prepare espresso drinks, cold brew, iced coffee, and drip coffee using commercial coffee equipment. The present disclosure also focuses on using computer vision to aid a robot in food and beverage making tasks so that these tasks may be completed even when the operational environment has and may continue to change. The present disclosure also allows a coffee-making solution where a human barista works alongside robots collaboratively.
In accordance with another aspect of the present disclosure, a robotic coffee preparation and serving station is provided that includes a six-axis robot arm controlled by executable software on a processor, gripper that can be controlled by the processor and a camera that is capable of taking high resolution images.
In accordance with another aspect of the present disclosure, a method of guiding a robot relative to an object that does not have a fixed location is provided. The method includes providing a camera, providing a visual cue associated with the object, detecting a pose of the visual cue with the camera, and providing to a controller a sequence of robot trajectory instructions relative to the pose of the visual cue to guide the robot to the object.
In accordance with another aspect of the foregoing disclosure, the method includes an initial step of pre-recording and saving in an electronic memory associated with the controller the sequence of robotic trajectory instructions.
In accordance with a further aspect of the foregoing disclosure, the method includes repeating the detecting and providing to the controller the sequence of robotic trajectory instructions until the desired food or beverage has been prepared and served.
In accordance with a further aspect of the foregoing disclosure, the method includes monitoring physical change in the environment and controlling the movement of the robot without colliding with the environment. The method includes providing a camera, providing a visual cue associated with the environment, detecting physical changes in the visual cue with the camera, and planning collision free robot trajectory accounting for the changes in the environment.
In accordance with a further aspect of the foregoing disclosure, the method includes monitoring various environmental conditions and guiding the robot to adjust to such environmental change. The method includes providing a camera, providing a visual cue associated with the object, detecting conditions of the visual cue with the camera, and adjusting the sequence of robot trajectory instructions to guide the robot handling the change in condition.
In accordance with a further aspect of the foregoing disclosure, the method includes identifying human beings and interacting with them. The method includes providing a camera, providing a visual cue associated with customers waiting to pick up their order, identifying who the order is for and guide the robot to deliver the order to the target customer.
The foregoing and other features and advantages of the present disclosure will be more readily appreciated as the same become better understood from the following detailed description when taken in conjunction with the accompanying drawings, wherein:
The present disclosure describes a system that allows one to capture complex motions to be executed by a robot, which are guided and modified by visual or other sensory input. Methods for such visual-guided systems are described below.
By using the camera 124, the robotic arm system 30 can place an object, such as a coffee cup, on a counter or other work surface without hitting existing objects, such as other cups, on the counter. After a food item or beverage is prepared, the robotic arm system 30 may place it on the counter or on another designated surface where the customer can pick it up. The robotic arm system 30 may use the on-board camera 124 to find an empty space to place the item so that it does not hit any objects on the counter. The computer vision system will also identify human hands and ensure moving parts, such as arm 126, avoids the human hands.
An example of the method shown in
A significant difference between the processes shown in
The invention of the present disclosure also provides ways to use computer vision to affect other aspects of coffee making. In
The invention of the present disclosure may use computer vision to handle dynamic changes in the environment. In
The present disclosure allows ways to use computer vision to identify customers waiting for their order, localizing them and controlling the robot to deliver the order to the target customer.
As desired, embodiments of the disclosure may include systems with more or fewer components than are illustrated in the drawings. Additionally, certain components of the systems may be combined in various embodiments of the disclosure. The systems described above are provided by way of example only.
The above description presents the best mode contemplated for carrying out the present embodiments, and of the manner and process of practicing them, in such full, clear, concise, and exact terms as to enable any person skilled in the art to which they pertain to practice these embodiments. The present embodiments are, however, susceptible to modifications and alternate constructions from those discussed above that are fully equivalent. Consequently, the present invention is not limited to the particular embodiments disclosed. On the contrary, the present invention covers all modifications and alternate constructions coming within the spirit and scope of the present disclosure. For example, the steps in the processes described herein need not be performed in the same order as they have been presented, and may be performed in any order(s). Further, steps that have been presented as being performed separately may in alternative embodiments be performed concurrently. Likewise, steps that have been presented as being performed concurrently may in alternative embodiments be performed separately.
The present application is a United States National Stage (§ 371) application of PCT/US21/33430 filed May 20, 2021, which claims the benefit of U.S. Provisional Patent Application Ser. No. 63/028,109, filed on May 21, 2020, the contents of which are hereby incorporated by reference in their entirety.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US21/33430 | 5/20/2021 | WO |
Number | Date | Country | |
---|---|---|---|
63028109 | May 2020 | US |